var/home/core/zuul-output/0000755000175000017500000000000015110237560014525 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015110252761015471 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005206050615110252752017701 0ustar rootrootNov 22 04:47:25 crc systemd[1]: Starting Kubernetes Kubelet... Nov 22 04:47:26 crc restorecon[4760]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:26 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 04:47:27 crc restorecon[4760]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 04:47:27 crc restorecon[4760]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 22 04:47:27 crc kubenswrapper[4818]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 22 04:47:27 crc kubenswrapper[4818]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 22 04:47:27 crc kubenswrapper[4818]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 22 04:47:27 crc kubenswrapper[4818]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 22 04:47:27 crc kubenswrapper[4818]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 22 04:47:27 crc kubenswrapper[4818]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.967687 4818 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979665 4818 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979700 4818 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979710 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979719 4818 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979728 4818 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979736 4818 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979745 4818 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979754 4818 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979762 4818 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979770 4818 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979778 4818 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979785 4818 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979793 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979800 4818 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979808 4818 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979816 4818 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979840 4818 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979851 4818 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979859 4818 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979867 4818 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979874 4818 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979882 4818 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979891 4818 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979899 4818 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979907 4818 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979917 4818 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979927 4818 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979935 4818 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979946 4818 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979956 4818 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979965 4818 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979973 4818 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979982 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.979993 4818 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980004 4818 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980012 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980020 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980028 4818 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980036 4818 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980044 4818 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980052 4818 feature_gate.go:330] unrecognized feature gate: Example Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980060 4818 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980067 4818 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980074 4818 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980082 4818 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980089 4818 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980097 4818 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980104 4818 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980113 4818 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980121 4818 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980129 4818 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980136 4818 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980144 4818 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980151 4818 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980158 4818 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980166 4818 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980174 4818 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980182 4818 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980190 4818 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980199 4818 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980208 4818 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980216 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980223 4818 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980231 4818 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980238 4818 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980246 4818 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980281 4818 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980289 4818 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980296 4818 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980304 4818 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.980311 4818 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980450 4818 flags.go:64] FLAG: --address="0.0.0.0" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980467 4818 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980481 4818 flags.go:64] FLAG: --anonymous-auth="true" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980492 4818 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980503 4818 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980513 4818 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980524 4818 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980535 4818 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980544 4818 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980553 4818 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980563 4818 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980572 4818 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980581 4818 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980591 4818 flags.go:64] FLAG: --cgroup-root="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980600 4818 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980609 4818 flags.go:64] FLAG: --client-ca-file="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980618 4818 flags.go:64] FLAG: --cloud-config="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980626 4818 flags.go:64] FLAG: --cloud-provider="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980635 4818 flags.go:64] FLAG: --cluster-dns="[]" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980645 4818 flags.go:64] FLAG: --cluster-domain="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980656 4818 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980666 4818 flags.go:64] FLAG: --config-dir="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980675 4818 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980684 4818 flags.go:64] FLAG: --container-log-max-files="5" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980694 4818 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980703 4818 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980713 4818 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980722 4818 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980731 4818 flags.go:64] FLAG: --contention-profiling="false" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980739 4818 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980748 4818 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980757 4818 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980766 4818 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980777 4818 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980785 4818 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980794 4818 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980803 4818 flags.go:64] FLAG: --enable-load-reader="false" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980812 4818 flags.go:64] FLAG: --enable-server="true" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980822 4818 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980833 4818 flags.go:64] FLAG: --event-burst="100" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980842 4818 flags.go:64] FLAG: --event-qps="50" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980851 4818 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980861 4818 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980870 4818 flags.go:64] FLAG: --eviction-hard="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980880 4818 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980889 4818 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980897 4818 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980907 4818 flags.go:64] FLAG: --eviction-soft="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980916 4818 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980926 4818 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980935 4818 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980944 4818 flags.go:64] FLAG: --experimental-mounter-path="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980953 4818 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980962 4818 flags.go:64] FLAG: --fail-swap-on="true" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980971 4818 flags.go:64] FLAG: --feature-gates="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980981 4818 flags.go:64] FLAG: --file-check-frequency="20s" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980990 4818 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.980999 4818 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981009 4818 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981017 4818 flags.go:64] FLAG: --healthz-port="10248" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981027 4818 flags.go:64] FLAG: --help="false" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981035 4818 flags.go:64] FLAG: --hostname-override="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981044 4818 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981053 4818 flags.go:64] FLAG: --http-check-frequency="20s" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981061 4818 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981070 4818 flags.go:64] FLAG: --image-credential-provider-config="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981079 4818 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981087 4818 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981097 4818 flags.go:64] FLAG: --image-service-endpoint="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981105 4818 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981114 4818 flags.go:64] FLAG: --kube-api-burst="100" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981123 4818 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981132 4818 flags.go:64] FLAG: --kube-api-qps="50" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981141 4818 flags.go:64] FLAG: --kube-reserved="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981150 4818 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981158 4818 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981167 4818 flags.go:64] FLAG: --kubelet-cgroups="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981177 4818 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981187 4818 flags.go:64] FLAG: --lock-file="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981197 4818 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981206 4818 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981215 4818 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981227 4818 flags.go:64] FLAG: --log-json-split-stream="false" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981236 4818 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981246 4818 flags.go:64] FLAG: --log-text-split-stream="false" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981280 4818 flags.go:64] FLAG: --logging-format="text" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981290 4818 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981299 4818 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981308 4818 flags.go:64] FLAG: --manifest-url="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981317 4818 flags.go:64] FLAG: --manifest-url-header="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981328 4818 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981338 4818 flags.go:64] FLAG: --max-open-files="1000000" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981349 4818 flags.go:64] FLAG: --max-pods="110" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981358 4818 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981367 4818 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981376 4818 flags.go:64] FLAG: --memory-manager-policy="None" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981385 4818 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981394 4818 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981402 4818 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981412 4818 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981430 4818 flags.go:64] FLAG: --node-status-max-images="50" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981439 4818 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981448 4818 flags.go:64] FLAG: --oom-score-adj="-999" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981457 4818 flags.go:64] FLAG: --pod-cidr="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981465 4818 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981479 4818 flags.go:64] FLAG: --pod-manifest-path="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981487 4818 flags.go:64] FLAG: --pod-max-pids="-1" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981496 4818 flags.go:64] FLAG: --pods-per-core="0" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981505 4818 flags.go:64] FLAG: --port="10250" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981514 4818 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981523 4818 flags.go:64] FLAG: --provider-id="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981531 4818 flags.go:64] FLAG: --qos-reserved="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981540 4818 flags.go:64] FLAG: --read-only-port="10255" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981549 4818 flags.go:64] FLAG: --register-node="true" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981558 4818 flags.go:64] FLAG: --register-schedulable="true" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981567 4818 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981580 4818 flags.go:64] FLAG: --registry-burst="10" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981589 4818 flags.go:64] FLAG: --registry-qps="5" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981598 4818 flags.go:64] FLAG: --reserved-cpus="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981606 4818 flags.go:64] FLAG: --reserved-memory="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981617 4818 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981627 4818 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981636 4818 flags.go:64] FLAG: --rotate-certificates="false" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981645 4818 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981654 4818 flags.go:64] FLAG: --runonce="false" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981663 4818 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981672 4818 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981681 4818 flags.go:64] FLAG: --seccomp-default="false" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981689 4818 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981698 4818 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981707 4818 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981716 4818 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981725 4818 flags.go:64] FLAG: --storage-driver-password="root" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981734 4818 flags.go:64] FLAG: --storage-driver-secure="false" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981743 4818 flags.go:64] FLAG: --storage-driver-table="stats" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981752 4818 flags.go:64] FLAG: --storage-driver-user="root" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981760 4818 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981769 4818 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981778 4818 flags.go:64] FLAG: --system-cgroups="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981786 4818 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981800 4818 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981809 4818 flags.go:64] FLAG: --tls-cert-file="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981818 4818 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981828 4818 flags.go:64] FLAG: --tls-min-version="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981836 4818 flags.go:64] FLAG: --tls-private-key-file="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981845 4818 flags.go:64] FLAG: --topology-manager-policy="none" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981854 4818 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981863 4818 flags.go:64] FLAG: --topology-manager-scope="container" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981873 4818 flags.go:64] FLAG: --v="2" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981885 4818 flags.go:64] FLAG: --version="false" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981929 4818 flags.go:64] FLAG: --vmodule="" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981940 4818 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.981950 4818 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982139 4818 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982150 4818 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982158 4818 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982169 4818 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982184 4818 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982192 4818 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982200 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982208 4818 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982216 4818 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982224 4818 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982234 4818 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982244 4818 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982276 4818 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982284 4818 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982296 4818 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982307 4818 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982316 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982324 4818 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982332 4818 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982340 4818 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982348 4818 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982355 4818 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982363 4818 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982371 4818 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982379 4818 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982387 4818 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982394 4818 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982402 4818 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982410 4818 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982418 4818 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982426 4818 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982434 4818 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982442 4818 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982449 4818 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982457 4818 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982465 4818 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982472 4818 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982480 4818 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982488 4818 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982495 4818 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982504 4818 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982512 4818 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982520 4818 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982528 4818 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982535 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982543 4818 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982551 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982561 4818 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982570 4818 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982580 4818 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982589 4818 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982598 4818 feature_gate.go:330] unrecognized feature gate: Example Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982606 4818 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982615 4818 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982624 4818 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982632 4818 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982641 4818 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982650 4818 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982658 4818 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982673 4818 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982682 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982694 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982703 4818 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982711 4818 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982718 4818 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982726 4818 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982734 4818 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982742 4818 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982750 4818 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982757 4818 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 22 04:47:27 crc kubenswrapper[4818]: W1122 04:47:27.982765 4818 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 22 04:47:27 crc kubenswrapper[4818]: I1122 04:47:27.985996 4818 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.000020 4818 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.000068 4818 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000219 4818 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000239 4818 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000287 4818 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000301 4818 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000311 4818 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000321 4818 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000332 4818 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000342 4818 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000352 4818 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000361 4818 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000370 4818 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000380 4818 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000389 4818 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000398 4818 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000407 4818 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000417 4818 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000430 4818 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000462 4818 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000474 4818 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000485 4818 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000495 4818 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000505 4818 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000516 4818 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000526 4818 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000536 4818 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000545 4818 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000555 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000565 4818 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000575 4818 feature_gate.go:330] unrecognized feature gate: Example Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000584 4818 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000594 4818 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000603 4818 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000615 4818 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000625 4818 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000635 4818 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000645 4818 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000655 4818 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000665 4818 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000676 4818 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000690 4818 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000704 4818 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000721 4818 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000731 4818 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000744 4818 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000756 4818 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000766 4818 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000777 4818 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000788 4818 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000799 4818 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000810 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000823 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000835 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000845 4818 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000856 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000866 4818 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000876 4818 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000886 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000895 4818 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000906 4818 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000916 4818 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000929 4818 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000944 4818 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000955 4818 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000966 4818 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000977 4818 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000988 4818 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.000999 4818 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001010 4818 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001020 4818 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001031 4818 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001040 4818 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.001057 4818 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001408 4818 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001431 4818 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001443 4818 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001454 4818 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001464 4818 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001476 4818 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001487 4818 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001498 4818 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001514 4818 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001529 4818 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001540 4818 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001550 4818 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001561 4818 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001572 4818 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001584 4818 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001593 4818 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001603 4818 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001614 4818 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001624 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001634 4818 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001644 4818 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001653 4818 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001664 4818 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001675 4818 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001684 4818 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001695 4818 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001704 4818 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001714 4818 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001725 4818 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001735 4818 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001745 4818 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001755 4818 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001765 4818 feature_gate.go:330] unrecognized feature gate: Example Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001779 4818 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001792 4818 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001803 4818 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001815 4818 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001825 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001837 4818 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001850 4818 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001863 4818 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001875 4818 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001886 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001940 4818 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001951 4818 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.001962 4818 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.002557 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.002580 4818 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.002785 4818 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.002803 4818 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.002870 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.002883 4818 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.002897 4818 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.002910 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.002924 4818 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.002935 4818 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.002946 4818 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.002958 4818 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.002969 4818 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.002981 4818 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.002993 4818 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.003004 4818 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.003015 4818 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.003025 4818 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.003035 4818 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.003045 4818 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.003056 4818 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.003066 4818 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.003077 4818 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.003087 4818 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.003097 4818 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.003114 4818 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.004512 4818 server.go:940] "Client rotation is on, will bootstrap in background" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.012016 4818 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.012227 4818 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.014498 4818 server.go:997] "Starting client certificate rotation" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.014555 4818 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.014822 4818 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-18 19:56:16.11325439 +0000 UTC Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.015036 4818 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.053181 4818 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.056093 4818 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 22 04:47:28 crc kubenswrapper[4818]: E1122 04:47:28.057230 4818 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.076638 4818 log.go:25] "Validated CRI v1 runtime API" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.123053 4818 log.go:25] "Validated CRI v1 image API" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.125805 4818 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.130876 4818 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-22-04-42-12-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.130921 4818 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.158512 4818 manager.go:217] Machine: {Timestamp:2025-11-22 04:47:28.155848736 +0000 UTC m=+0.730265323 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:e865ab18-bc50-41fa-b6be-f0543f192f64 BootID:64c9fbbd-60c5-409c-a987-08eb342f18a6 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:12:14:19 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:12:14:19 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:b1:e7:f1 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:66:df:ac Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:96:fc:da Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:81:23:0c Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:1e:36:00 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:fe:5e:61:e5:e6:41 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:da:bb:dd:cd:1d:ca Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.159326 4818 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.159651 4818 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.161434 4818 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.161970 4818 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.162142 4818 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.162636 4818 topology_manager.go:138] "Creating topology manager with none policy" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.162750 4818 container_manager_linux.go:303] "Creating device plugin manager" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.163476 4818 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.163667 4818 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.164207 4818 state_mem.go:36] "Initialized new in-memory state store" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.164584 4818 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.170125 4818 kubelet.go:418] "Attempting to sync node with API server" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.170347 4818 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.170494 4818 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.170633 4818 kubelet.go:324] "Adding apiserver pod source" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.170789 4818 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.176358 4818 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.177840 4818 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.178348 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 22 04:47:28 crc kubenswrapper[4818]: E1122 04:47:28.178448 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.178477 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 22 04:47:28 crc kubenswrapper[4818]: E1122 04:47:28.178586 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.180472 4818 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.184369 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.184423 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.184443 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.184461 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.184488 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.184506 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.184544 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.184573 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.184594 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.184611 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.184661 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.184675 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.184724 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.185454 4818 server.go:1280] "Started kubelet" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.185590 4818 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.186013 4818 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.185833 4818 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.186955 4818 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 22 04:47:28 crc systemd[1]: Started Kubernetes Kubelet. Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.195784 4818 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.195844 4818 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.196384 4818 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.196344 4818 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 05:06:23.121881202 +0000 UTC Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.196444 4818 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.196455 4818 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 576h18m54.925431555s for next certificate rotation Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.196433 4818 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 22 04:47:28 crc kubenswrapper[4818]: E1122 04:47:28.196550 4818 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.197621 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 22 04:47:28 crc kubenswrapper[4818]: E1122 04:47:28.197747 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.198304 4818 factory.go:55] Registering systemd factory Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.198348 4818 factory.go:221] Registration of the systemd container factory successfully Nov 22 04:47:28 crc kubenswrapper[4818]: E1122 04:47:28.199527 4818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.110:6443: connect: connection refused" interval="200ms" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.199689 4818 factory.go:153] Registering CRI-O factory Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.199719 4818 factory.go:221] Registration of the crio container factory successfully Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.199890 4818 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.199950 4818 factory.go:103] Registering Raw factory Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.199974 4818 manager.go:1196] Started watching for new ooms in manager Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.201034 4818 manager.go:319] Starting recovery of all containers Nov 22 04:47:28 crc kubenswrapper[4818]: E1122 04:47:28.219347 4818 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.110:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187a3acb2198f2de default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-22 04:47:28.185414366 +0000 UTC m=+0.759830923,LastTimestamp:2025-11-22 04:47:28.185414366 +0000 UTC m=+0.759830923,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.244449 4818 server.go:460] "Adding debug handlers to kubelet server" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.253497 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.253911 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.253933 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.253954 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.253974 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.253991 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254009 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254026 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254046 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254065 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254083 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254101 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254120 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254141 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254158 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254181 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254201 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254218 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254236 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254288 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254315 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254333 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254353 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254371 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254390 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254407 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254430 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254448 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254466 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254483 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254499 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254519 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254543 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254560 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254577 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254593 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254611 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254628 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254646 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254663 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254681 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254698 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254716 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254734 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254752 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254769 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254787 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254809 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254827 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254845 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254863 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254879 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254901 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254921 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254940 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254972 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.254990 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255007 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255026 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255043 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255059 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255076 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255098 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255122 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255151 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255176 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255201 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255223 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255246 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255303 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255328 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255350 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255381 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255403 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255426 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255448 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255470 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255491 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255515 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255539 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255562 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255583 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255605 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255628 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255651 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255675 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255701 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255726 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255752 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255776 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.255803 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.257973 4818 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258093 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258138 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258163 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258185 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258204 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258224 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258244 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258291 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258309 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258331 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258349 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258367 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258386 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258415 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258442 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258467 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258490 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258514 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258548 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258568 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258589 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258610 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258629 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258647 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258666 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258686 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258703 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258721 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258742 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258761 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258779 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258798 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258817 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258834 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258859 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258879 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258897 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258913 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258934 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258951 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258970 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.258989 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259006 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259024 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259041 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259060 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259078 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259096 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259114 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259132 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259151 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259169 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259190 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259208 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259226 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259243 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259292 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259312 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259329 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259348 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259366 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259386 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259403 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259420 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259440 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259457 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259477 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259495 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259515 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259532 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259552 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259569 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259587 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259607 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259624 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259643 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259660 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259678 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259698 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259716 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259734 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259754 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259773 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259791 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259810 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259828 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259856 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259874 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259892 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259911 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259930 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259948 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259965 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.259985 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260003 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260021 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260040 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260059 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260079 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260097 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260115 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260135 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260153 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260172 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260192 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260211 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260230 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260248 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260293 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260310 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260328 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260346 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260363 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260382 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260401 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260419 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260437 4818 reconstruct.go:97] "Volume reconstruction finished" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.260449 4818 reconciler.go:26] "Reconciler: start to sync state" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.263857 4818 manager.go:324] Recovery completed Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.280910 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.283106 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.283175 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.283197 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.286729 4818 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.286769 4818 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.286798 4818 state_mem.go:36] "Initialized new in-memory state store" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.286695 4818 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.289733 4818 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.289797 4818 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.289861 4818 kubelet.go:2335] "Starting kubelet main sync loop" Nov 22 04:47:28 crc kubenswrapper[4818]: E1122 04:47:28.289934 4818 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.290728 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 22 04:47:28 crc kubenswrapper[4818]: E1122 04:47:28.290790 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 22 04:47:28 crc kubenswrapper[4818]: E1122 04:47:28.297050 4818 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.311653 4818 policy_none.go:49] "None policy: Start" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.312726 4818 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.312898 4818 state_mem.go:35] "Initializing new in-memory state store" Nov 22 04:47:28 crc kubenswrapper[4818]: E1122 04:47:28.390332 4818 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.394865 4818 manager.go:334] "Starting Device Plugin manager" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.395113 4818 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.395145 4818 server.go:79] "Starting device plugin registration server" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.395532 4818 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.395553 4818 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.395945 4818 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.396093 4818 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.396103 4818 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 22 04:47:28 crc kubenswrapper[4818]: E1122 04:47:28.400116 4818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.110:6443: connect: connection refused" interval="400ms" Nov 22 04:47:28 crc kubenswrapper[4818]: E1122 04:47:28.403165 4818 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.497163 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.498898 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.498939 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.498950 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.498979 4818 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 04:47:28 crc kubenswrapper[4818]: E1122 04:47:28.499569 4818 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.110:6443: connect: connection refused" node="crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.590757 4818 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.590909 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.592686 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.592726 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.592735 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.592868 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.593056 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.593108 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.593624 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.593644 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.593654 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.593759 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.593935 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.593978 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.594110 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.594180 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.594237 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.594546 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.594581 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.594597 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.594739 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.594965 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.595052 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.594998 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.595161 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.595175 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.595796 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.595835 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.595851 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.595983 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.596178 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.596232 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.597167 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.597195 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.597196 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.597243 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.597312 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.597212 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.597375 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.597401 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.597205 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.597620 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.597685 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.598667 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.598706 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.598722 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.665005 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.665119 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.665152 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.665186 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.665213 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.665282 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.665313 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.665341 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.665371 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.665399 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.665431 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.665472 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.665518 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.665547 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.665573 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.700281 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.701530 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.701570 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.701579 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.701600 4818 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 04:47:28 crc kubenswrapper[4818]: E1122 04:47:28.702033 4818 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.110:6443: connect: connection refused" node="crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.767497 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.767571 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.767597 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.767619 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.767639 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.767661 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.767722 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.767744 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.767766 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.767788 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.767811 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.767832 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.767853 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.767854 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.767926 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.767927 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.767874 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.768011 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.768028 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.768045 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.768084 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.768121 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.767853 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.768157 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.768203 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.768227 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.768238 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.768108 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.768329 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.768350 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: E1122 04:47:28.801592 4818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.110:6443: connect: connection refused" interval="800ms" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.932893 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.953140 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.967496 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.989589 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.992831 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-c472df278ad85a57d98d89daa02e675d704986b0d34f9e8a1ba3179e0afb8661 WatchSource:0}: Error finding container c472df278ad85a57d98d89daa02e675d704986b0d34f9e8a1ba3179e0afb8661: Status 404 returned error can't find the container with id c472df278ad85a57d98d89daa02e675d704986b0d34f9e8a1ba3179e0afb8661 Nov 22 04:47:28 crc kubenswrapper[4818]: W1122 04:47:28.994710 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-0186279d571116a2d40d5251b6387140a63e80730637afb1e3b04bdeb0fea2c9 WatchSource:0}: Error finding container 0186279d571116a2d40d5251b6387140a63e80730637afb1e3b04bdeb0fea2c9: Status 404 returned error can't find the container with id 0186279d571116a2d40d5251b6387140a63e80730637afb1e3b04bdeb0fea2c9 Nov 22 04:47:28 crc kubenswrapper[4818]: I1122 04:47:28.995826 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 04:47:29 crc kubenswrapper[4818]: W1122 04:47:29.004119 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-f9ff987de3cc9b28166179eb6cc5c4bb436f1631281492a534d1e7e0caf5d1e5 WatchSource:0}: Error finding container f9ff987de3cc9b28166179eb6cc5c4bb436f1631281492a534d1e7e0caf5d1e5: Status 404 returned error can't find the container with id f9ff987de3cc9b28166179eb6cc5c4bb436f1631281492a534d1e7e0caf5d1e5 Nov 22 04:47:29 crc kubenswrapper[4818]: W1122 04:47:29.013481 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-442f1e9cdbf7344786798e9acf950a165d89d81f570d0f5bb7d36cd76e533da8 WatchSource:0}: Error finding container 442f1e9cdbf7344786798e9acf950a165d89d81f570d0f5bb7d36cd76e533da8: Status 404 returned error can't find the container with id 442f1e9cdbf7344786798e9acf950a165d89d81f570d0f5bb7d36cd76e533da8 Nov 22 04:47:29 crc kubenswrapper[4818]: W1122 04:47:29.027290 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-76db6c7810cc794b7048fd7cb67b02beec9866cebe326665aa69e5eb4050743a WatchSource:0}: Error finding container 76db6c7810cc794b7048fd7cb67b02beec9866cebe326665aa69e5eb4050743a: Status 404 returned error can't find the container with id 76db6c7810cc794b7048fd7cb67b02beec9866cebe326665aa69e5eb4050743a Nov 22 04:47:29 crc kubenswrapper[4818]: I1122 04:47:29.102625 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:29 crc kubenswrapper[4818]: I1122 04:47:29.104737 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:29 crc kubenswrapper[4818]: I1122 04:47:29.104791 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:29 crc kubenswrapper[4818]: I1122 04:47:29.104847 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:29 crc kubenswrapper[4818]: I1122 04:47:29.104881 4818 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 04:47:29 crc kubenswrapper[4818]: E1122 04:47:29.105518 4818 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.110:6443: connect: connection refused" node="crc" Nov 22 04:47:29 crc kubenswrapper[4818]: W1122 04:47:29.137530 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 22 04:47:29 crc kubenswrapper[4818]: E1122 04:47:29.137708 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 22 04:47:29 crc kubenswrapper[4818]: I1122 04:47:29.187199 4818 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 22 04:47:29 crc kubenswrapper[4818]: W1122 04:47:29.193944 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 22 04:47:29 crc kubenswrapper[4818]: E1122 04:47:29.194087 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 22 04:47:29 crc kubenswrapper[4818]: I1122 04:47:29.295657 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"76db6c7810cc794b7048fd7cb67b02beec9866cebe326665aa69e5eb4050743a"} Nov 22 04:47:29 crc kubenswrapper[4818]: I1122 04:47:29.297795 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"442f1e9cdbf7344786798e9acf950a165d89d81f570d0f5bb7d36cd76e533da8"} Nov 22 04:47:29 crc kubenswrapper[4818]: I1122 04:47:29.300079 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f9ff987de3cc9b28166179eb6cc5c4bb436f1631281492a534d1e7e0caf5d1e5"} Nov 22 04:47:29 crc kubenswrapper[4818]: I1122 04:47:29.301346 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c472df278ad85a57d98d89daa02e675d704986b0d34f9e8a1ba3179e0afb8661"} Nov 22 04:47:29 crc kubenswrapper[4818]: I1122 04:47:29.308210 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"0186279d571116a2d40d5251b6387140a63e80730637afb1e3b04bdeb0fea2c9"} Nov 22 04:47:29 crc kubenswrapper[4818]: W1122 04:47:29.317732 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 22 04:47:29 crc kubenswrapper[4818]: E1122 04:47:29.317822 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 22 04:47:29 crc kubenswrapper[4818]: W1122 04:47:29.486526 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 22 04:47:29 crc kubenswrapper[4818]: E1122 04:47:29.486633 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 22 04:47:29 crc kubenswrapper[4818]: E1122 04:47:29.602601 4818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.110:6443: connect: connection refused" interval="1.6s" Nov 22 04:47:29 crc kubenswrapper[4818]: I1122 04:47:29.906376 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:29 crc kubenswrapper[4818]: I1122 04:47:29.908054 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:29 crc kubenswrapper[4818]: I1122 04:47:29.908117 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:29 crc kubenswrapper[4818]: I1122 04:47:29.908140 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:29 crc kubenswrapper[4818]: I1122 04:47:29.908181 4818 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 04:47:29 crc kubenswrapper[4818]: E1122 04:47:29.908861 4818 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.110:6443: connect: connection refused" node="crc" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.186657 4818 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.238834 4818 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 22 04:47:30 crc kubenswrapper[4818]: E1122 04:47:30.239872 4818 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.313217 4818 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a" exitCode=0 Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.313383 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a"} Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.313403 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.314685 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.314734 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.314751 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.315546 4818 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ccc8845abe9f2530b18a2aba39b064bb0f7bdfb31dfa1331acb6499156db4ea0" exitCode=0 Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.315641 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ccc8845abe9f2530b18a2aba39b064bb0f7bdfb31dfa1331acb6499156db4ea0"} Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.315673 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.316775 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.316847 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.316874 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.318842 4818 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="bab5d050214f11d3f38e205479fd500d3ef975f33f9026a5a49d9eb41d400cb9" exitCode=0 Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.318906 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"bab5d050214f11d3f38e205479fd500d3ef975f33f9026a5a49d9eb41d400cb9"} Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.318945 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.319912 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.320197 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.320218 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.320228 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.321278 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.321327 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.321350 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.321987 4818 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00" exitCode=0 Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.322062 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00"} Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.322092 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.324691 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.324739 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.324760 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.325970 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1b13b8162d6f7654a176f320274869a5df27e278f935f0cab1068403f9e12f8b"} Nov 22 04:47:30 crc kubenswrapper[4818]: I1122 04:47:30.326012 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"152036fde48e3fdcd33619dc1ca10977d41bb97c16bca4145926b4afe17267b9"} Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.186831 4818 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 22 04:47:31 crc kubenswrapper[4818]: E1122 04:47:31.204362 4818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.110:6443: connect: connection refused" interval="3.2s" Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.331714 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8e315f3e5e87e5ed47280e8efc909148c1b18af404778609c867f572a1c352c9"} Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.331790 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"976c49c2e32cd03a00e6a0a91e325864c9a1eb48cc412cc03ed6cdd620da9125"} Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.336228 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5690052f421eefbb317162b2ab5344182999498110d1bd671cc11116f4a8256b"} Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.336340 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.336361 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e58d44809fc7aff83cf6c883f2db88130b98afdd8d29ea1a482560616853c207"} Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.337687 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.337728 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.337742 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.339766 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4"} Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.339808 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce"} Nov 22 04:47:31 crc kubenswrapper[4818]: W1122 04:47:31.339786 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 22 04:47:31 crc kubenswrapper[4818]: E1122 04:47:31.339880 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.343412 4818 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="41515325c6456c08b79496fe17f63e7670d56ed4dc61ed276fb51b460ab5e87e" exitCode=0 Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.343482 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"41515325c6456c08b79496fe17f63e7670d56ed4dc61ed276fb51b460ab5e87e"} Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.343573 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.345857 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.345927 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.345953 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.348225 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"9682e4f09980ee808df4b61b0bab8d8dd9d713a6211e5b3b6cf4ae6dffc84aa8"} Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.348340 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.350178 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.350237 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.350295 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.508976 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.509945 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.509969 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.509980 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:31 crc kubenswrapper[4818]: I1122 04:47:31.509999 4818 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 04:47:31 crc kubenswrapper[4818]: E1122 04:47:31.510407 4818 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.110:6443: connect: connection refused" node="crc" Nov 22 04:47:31 crc kubenswrapper[4818]: W1122 04:47:31.799566 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 22 04:47:31 crc kubenswrapper[4818]: E1122 04:47:31.799703 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.187523 4818 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 22 04:47:32 crc kubenswrapper[4818]: W1122 04:47:32.213306 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 22 04:47:32 crc kubenswrapper[4818]: E1122 04:47:32.213420 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 22 04:47:32 crc kubenswrapper[4818]: W1122 04:47:32.256233 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 22 04:47:32 crc kubenswrapper[4818]: E1122 04:47:32.256415 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.353896 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"fe9de7b662c0ba4d848448a8322d3a464e58337895bf25390d4df018b695834a"} Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.354076 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.355224 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.355295 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.355312 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.360144 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c1dcee3f73772f0f98d8f5f4a900f23ac0c53b9fd0ab37d6c61995a3135713a7"} Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.360201 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0"} Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.360229 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4"} Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.360395 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.361956 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.361995 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.362011 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.362939 4818 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ddaa1fc41d78fa6bd011a178127d48fbec3859541b81be1c1fdde382e2777135" exitCode=0 Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.363126 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.363321 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.363185 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.362983 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ddaa1fc41d78fa6bd011a178127d48fbec3859541b81be1c1fdde382e2777135"} Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.364494 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.364529 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.364547 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.364822 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.365000 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.365123 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.365179 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.365205 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:32 crc kubenswrapper[4818]: I1122 04:47:32.365153 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:33 crc kubenswrapper[4818]: I1122 04:47:33.186758 4818 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 22 04:47:33 crc kubenswrapper[4818]: I1122 04:47:33.367362 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0aadde793d0cd469559cedaa5d12ca8538604e9e400330380fda799d5abba734"} Nov 22 04:47:33 crc kubenswrapper[4818]: I1122 04:47:33.367422 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 04:47:33 crc kubenswrapper[4818]: I1122 04:47:33.367461 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:33 crc kubenswrapper[4818]: I1122 04:47:33.367479 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 04:47:33 crc kubenswrapper[4818]: I1122 04:47:33.367531 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:33 crc kubenswrapper[4818]: I1122 04:47:33.368457 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:33 crc kubenswrapper[4818]: I1122 04:47:33.368514 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:33 crc kubenswrapper[4818]: I1122 04:47:33.368588 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:33 crc kubenswrapper[4818]: I1122 04:47:33.368599 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:33 crc kubenswrapper[4818]: I1122 04:47:33.368563 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:33 crc kubenswrapper[4818]: I1122 04:47:33.368678 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:33 crc kubenswrapper[4818]: I1122 04:47:33.658629 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 04:47:33 crc kubenswrapper[4818]: E1122 04:47:33.813900 4818 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.110:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187a3acb2198f2de default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-22 04:47:28.185414366 +0000 UTC m=+0.759830923,LastTimestamp:2025-11-22 04:47:28.185414366 +0000 UTC m=+0.759830923,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 22 04:47:33 crc kubenswrapper[4818]: I1122 04:47:33.841539 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 04:47:33 crc kubenswrapper[4818]: I1122 04:47:33.871910 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 04:47:33 crc kubenswrapper[4818]: I1122 04:47:33.872134 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:33 crc kubenswrapper[4818]: I1122 04:47:33.873555 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:33 crc kubenswrapper[4818]: I1122 04:47:33.873627 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:33 crc kubenswrapper[4818]: I1122 04:47:33.873650 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.187030 4818 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.350347 4818 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 22 04:47:34 crc kubenswrapper[4818]: E1122 04:47:34.351585 4818 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.372815 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.376001 4818 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c1dcee3f73772f0f98d8f5f4a900f23ac0c53b9fd0ab37d6c61995a3135713a7" exitCode=255 Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.376089 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"c1dcee3f73772f0f98d8f5f4a900f23ac0c53b9fd0ab37d6c61995a3135713a7"} Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.376174 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.377609 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.377644 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.377654 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.378245 4818 scope.go:117] "RemoveContainer" containerID="c1dcee3f73772f0f98d8f5f4a900f23ac0c53b9fd0ab37d6c61995a3135713a7" Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.384846 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.385216 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b0d5fa59c436fa3b15782740e4c52b28f161dee88f7066780b0600abbac26871"} Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.385339 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"497c0b377c13ff186c85a1a98818a8aeb3adf0c705ae977e968bc8933b137934"} Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.389401 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.389475 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.389502 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:34 crc kubenswrapper[4818]: E1122 04:47:34.405204 4818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.110:6443: connect: connection refused" interval="6.4s" Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.710676 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.712504 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.712691 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.712728 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:34 crc kubenswrapper[4818]: I1122 04:47:34.712776 4818 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 04:47:35 crc kubenswrapper[4818]: I1122 04:47:35.100450 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 04:47:35 crc kubenswrapper[4818]: I1122 04:47:35.150678 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 04:47:35 crc kubenswrapper[4818]: I1122 04:47:35.150898 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:35 crc kubenswrapper[4818]: I1122 04:47:35.152738 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:35 crc kubenswrapper[4818]: I1122 04:47:35.152775 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:35 crc kubenswrapper[4818]: I1122 04:47:35.152793 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:35 crc kubenswrapper[4818]: I1122 04:47:35.317043 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 04:47:35 crc kubenswrapper[4818]: I1122 04:47:35.390686 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 22 04:47:35 crc kubenswrapper[4818]: I1122 04:47:35.393370 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe"} Nov 22 04:47:35 crc kubenswrapper[4818]: I1122 04:47:35.393503 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:35 crc kubenswrapper[4818]: I1122 04:47:35.396591 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:35 crc kubenswrapper[4818]: I1122 04:47:35.396636 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:35 crc kubenswrapper[4818]: I1122 04:47:35.396653 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:35 crc kubenswrapper[4818]: I1122 04:47:35.400838 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"cc48d49c0d054bc2d6af8010f53b799125ad4b3e67290d5c8da84b5a4a516e36"} Nov 22 04:47:35 crc kubenswrapper[4818]: I1122 04:47:35.400897 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8a760ca1879dd0698f20fe1a63f632243280be797c01ac221cd5b17dfd54f415"} Nov 22 04:47:35 crc kubenswrapper[4818]: I1122 04:47:35.401006 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:35 crc kubenswrapper[4818]: I1122 04:47:35.402209 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:35 crc kubenswrapper[4818]: I1122 04:47:35.402305 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:35 crc kubenswrapper[4818]: I1122 04:47:35.402334 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:36 crc kubenswrapper[4818]: I1122 04:47:36.403197 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:36 crc kubenswrapper[4818]: I1122 04:47:36.403231 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 04:47:36 crc kubenswrapper[4818]: I1122 04:47:36.403198 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:36 crc kubenswrapper[4818]: I1122 04:47:36.404699 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:36 crc kubenswrapper[4818]: I1122 04:47:36.404752 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:36 crc kubenswrapper[4818]: I1122 04:47:36.404817 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:36 crc kubenswrapper[4818]: I1122 04:47:36.404843 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:36 crc kubenswrapper[4818]: I1122 04:47:36.404764 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:36 crc kubenswrapper[4818]: I1122 04:47:36.404932 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:36 crc kubenswrapper[4818]: I1122 04:47:36.645999 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 22 04:47:37 crc kubenswrapper[4818]: I1122 04:47:37.406054 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:37 crc kubenswrapper[4818]: I1122 04:47:37.406110 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:37 crc kubenswrapper[4818]: I1122 04:47:37.408647 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:37 crc kubenswrapper[4818]: I1122 04:47:37.408681 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:37 crc kubenswrapper[4818]: I1122 04:47:37.408725 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:37 crc kubenswrapper[4818]: I1122 04:47:37.408750 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:37 crc kubenswrapper[4818]: I1122 04:47:37.408695 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:37 crc kubenswrapper[4818]: I1122 04:47:37.408825 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:37 crc kubenswrapper[4818]: I1122 04:47:37.518083 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 22 04:47:38 crc kubenswrapper[4818]: E1122 04:47:38.403358 4818 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 22 04:47:38 crc kubenswrapper[4818]: I1122 04:47:38.407601 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:38 crc kubenswrapper[4818]: I1122 04:47:38.409207 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:38 crc kubenswrapper[4818]: I1122 04:47:38.409249 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:38 crc kubenswrapper[4818]: I1122 04:47:38.409287 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:39 crc kubenswrapper[4818]: I1122 04:47:39.722518 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 04:47:39 crc kubenswrapper[4818]: I1122 04:47:39.725539 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:39 crc kubenswrapper[4818]: I1122 04:47:39.727808 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:39 crc kubenswrapper[4818]: I1122 04:47:39.727852 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:39 crc kubenswrapper[4818]: I1122 04:47:39.727870 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:39 crc kubenswrapper[4818]: I1122 04:47:39.731803 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 04:47:39 crc kubenswrapper[4818]: I1122 04:47:39.834279 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 04:47:40 crc kubenswrapper[4818]: I1122 04:47:40.413873 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:40 crc kubenswrapper[4818]: I1122 04:47:40.415413 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:40 crc kubenswrapper[4818]: I1122 04:47:40.415502 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:40 crc kubenswrapper[4818]: I1122 04:47:40.415522 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:40 crc kubenswrapper[4818]: I1122 04:47:40.420390 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 04:47:41 crc kubenswrapper[4818]: I1122 04:47:41.417452 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:41 crc kubenswrapper[4818]: I1122 04:47:41.418892 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:41 crc kubenswrapper[4818]: I1122 04:47:41.418998 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:41 crc kubenswrapper[4818]: I1122 04:47:41.419085 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:42 crc kubenswrapper[4818]: I1122 04:47:42.419620 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:42 crc kubenswrapper[4818]: I1122 04:47:42.421059 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:42 crc kubenswrapper[4818]: I1122 04:47:42.421142 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:42 crc kubenswrapper[4818]: I1122 04:47:42.421170 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:42 crc kubenswrapper[4818]: I1122 04:47:42.835107 4818 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 04:47:42 crc kubenswrapper[4818]: I1122 04:47:42.835198 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 22 04:47:42 crc kubenswrapper[4818]: I1122 04:47:42.930905 4818 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 22 04:47:44 crc kubenswrapper[4818]: E1122 04:47:44.714284 4818 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Nov 22 04:47:45 crc kubenswrapper[4818]: I1122 04:47:45.188330 4818 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 22 04:47:45 crc kubenswrapper[4818]: I1122 04:47:45.317271 4818 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 04:47:45 crc kubenswrapper[4818]: I1122 04:47:45.317335 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 22 04:47:45 crc kubenswrapper[4818]: W1122 04:47:45.577025 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 22 04:47:45 crc kubenswrapper[4818]: I1122 04:47:45.577149 4818 trace.go:236] Trace[1653231459]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Nov-2025 04:47:35.575) (total time: 10001ms): Nov 22 04:47:45 crc kubenswrapper[4818]: Trace[1653231459]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (04:47:45.577) Nov 22 04:47:45 crc kubenswrapper[4818]: Trace[1653231459]: [10.001914241s] [10.001914241s] END Nov 22 04:47:45 crc kubenswrapper[4818]: E1122 04:47:45.577186 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 22 04:47:45 crc kubenswrapper[4818]: I1122 04:47:45.918197 4818 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 22 04:47:45 crc kubenswrapper[4818]: I1122 04:47:45.918865 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 22 04:47:47 crc kubenswrapper[4818]: I1122 04:47:47.587910 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 22 04:47:47 crc kubenswrapper[4818]: I1122 04:47:47.588234 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:47 crc kubenswrapper[4818]: I1122 04:47:47.589724 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:47 crc kubenswrapper[4818]: I1122 04:47:47.589778 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:47 crc kubenswrapper[4818]: I1122 04:47:47.589799 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:47 crc kubenswrapper[4818]: I1122 04:47:47.606555 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 22 04:47:48 crc kubenswrapper[4818]: E1122 04:47:48.403556 4818 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 22 04:47:48 crc kubenswrapper[4818]: I1122 04:47:48.437766 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:48 crc kubenswrapper[4818]: I1122 04:47:48.438683 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:48 crc kubenswrapper[4818]: I1122 04:47:48.438719 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:48 crc kubenswrapper[4818]: I1122 04:47:48.438730 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:50 crc kubenswrapper[4818]: I1122 04:47:50.324817 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 04:47:50 crc kubenswrapper[4818]: I1122 04:47:50.325020 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:50 crc kubenswrapper[4818]: I1122 04:47:50.327046 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:50 crc kubenswrapper[4818]: I1122 04:47:50.327078 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:50 crc kubenswrapper[4818]: I1122 04:47:50.327088 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:50 crc kubenswrapper[4818]: I1122 04:47:50.332670 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 04:47:50 crc kubenswrapper[4818]: I1122 04:47:50.443596 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:50 crc kubenswrapper[4818]: I1122 04:47:50.445085 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:50 crc kubenswrapper[4818]: I1122 04:47:50.445138 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:50 crc kubenswrapper[4818]: I1122 04:47:50.445155 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:50 crc kubenswrapper[4818]: E1122 04:47:50.921438 4818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="7s" Nov 22 04:47:50 crc kubenswrapper[4818]: I1122 04:47:50.923726 4818 trace.go:236] Trace[1110166036]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Nov-2025 04:47:36.166) (total time: 14757ms): Nov 22 04:47:50 crc kubenswrapper[4818]: Trace[1110166036]: ---"Objects listed" error: 14757ms (04:47:50.923) Nov 22 04:47:50 crc kubenswrapper[4818]: Trace[1110166036]: [14.757435398s] [14.757435398s] END Nov 22 04:47:50 crc kubenswrapper[4818]: I1122 04:47:50.923768 4818 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 22 04:47:50 crc kubenswrapper[4818]: I1122 04:47:50.943986 4818 trace.go:236] Trace[1128329349]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Nov-2025 04:47:38.241) (total time: 12702ms): Nov 22 04:47:50 crc kubenswrapper[4818]: Trace[1128329349]: ---"Objects listed" error: 12702ms (04:47:50.943) Nov 22 04:47:50 crc kubenswrapper[4818]: Trace[1128329349]: [12.702386816s] [12.702386816s] END Nov 22 04:47:50 crc kubenswrapper[4818]: I1122 04:47:50.944036 4818 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 22 04:47:50 crc kubenswrapper[4818]: I1122 04:47:50.951386 4818 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 22 04:47:50 crc kubenswrapper[4818]: I1122 04:47:50.954986 4818 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Nov 22 04:47:50 crc kubenswrapper[4818]: I1122 04:47:50.956531 4818 trace.go:236] Trace[2078044473]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Nov-2025 04:47:37.893) (total time: 13063ms): Nov 22 04:47:50 crc kubenswrapper[4818]: Trace[2078044473]: ---"Objects listed" error: 13063ms (04:47:50.956) Nov 22 04:47:50 crc kubenswrapper[4818]: Trace[2078044473]: [13.063193908s] [13.063193908s] END Nov 22 04:47:50 crc kubenswrapper[4818]: I1122 04:47:50.956568 4818 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.000640 4818 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:35782->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.000718 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:35782->192.168.126.11:17697: read: connection reset by peer" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.000652 4818 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:35792->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.000868 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:35792->192.168.126.11:17697: read: connection reset by peer" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.001439 4818 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.001494 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.114425 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.115614 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.115648 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.115659 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.115812 4818 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.131364 4818 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.131592 4818 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 22 04:47:51 crc kubenswrapper[4818]: E1122 04:47:51.131609 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.134307 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.134348 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.134358 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.134375 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.134385 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:51Z","lastTransitionTime":"2025-11-22T04:47:51Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Nov 22 04:47:51 crc kubenswrapper[4818]: E1122 04:47:51.145028 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.148497 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.148533 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.148542 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.148562 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.148572 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:51Z","lastTransitionTime":"2025-11-22T04:47:51Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Nov 22 04:47:51 crc kubenswrapper[4818]: E1122 04:47:51.157601 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.160920 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.160955 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.160964 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.160980 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.160996 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:51Z","lastTransitionTime":"2025-11-22T04:47:51Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Nov 22 04:47:51 crc kubenswrapper[4818]: E1122 04:47:51.174232 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.177101 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.177135 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.177144 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.177160 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.177169 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:51Z","lastTransitionTime":"2025-11-22T04:47:51Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Nov 22 04:47:51 crc kubenswrapper[4818]: E1122 04:47:51.187513 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:51 crc kubenswrapper[4818]: E1122 04:47:51.187636 4818 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 04:47:51 crc kubenswrapper[4818]: E1122 04:47:51.187658 4818 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 04:47:51 crc kubenswrapper[4818]: E1122 04:47:51.288416 4818 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 04:47:51 crc kubenswrapper[4818]: E1122 04:47:51.389229 4818 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.448116 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.448902 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.450983 4818 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe" exitCode=255 Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.451025 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe"} Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.451081 4818 scope.go:117] "RemoveContainer" containerID="c1dcee3f73772f0f98d8f5f4a900f23ac0c53b9fd0ab37d6c61995a3135713a7" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.451273 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.452191 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.452236 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.452269 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.452981 4818 scope.go:117] "RemoveContainer" containerID="e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe" Nov 22 04:47:51 crc kubenswrapper[4818]: E1122 04:47:51.453188 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 22 04:47:51 crc kubenswrapper[4818]: E1122 04:47:51.490100 4818 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 04:47:51 crc kubenswrapper[4818]: E1122 04:47:51.590824 4818 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.673943 4818 csr.go:261] certificate signing request csr-tw7jj is approved, waiting to be issued Nov 22 04:47:51 crc kubenswrapper[4818]: E1122 04:47:51.691210 4818 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.731139 4818 csr.go:257] certificate signing request csr-tw7jj is issued Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.776359 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.776488 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.777466 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.777504 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.777513 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:51 crc kubenswrapper[4818]: I1122 04:47:51.782350 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 04:47:51 crc kubenswrapper[4818]: E1122 04:47:51.791577 4818 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 04:47:51 crc kubenswrapper[4818]: E1122 04:47:51.892128 4818 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 04:47:51 crc kubenswrapper[4818]: E1122 04:47:51.992663 4818 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 04:47:52 crc kubenswrapper[4818]: E1122 04:47:52.093620 4818 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 04:47:52 crc kubenswrapper[4818]: E1122 04:47:52.194652 4818 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 04:47:52 crc kubenswrapper[4818]: E1122 04:47:52.294759 4818 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 04:47:52 crc kubenswrapper[4818]: E1122 04:47:52.395871 4818 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.454147 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.455371 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.455937 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.455961 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.455969 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:52 crc kubenswrapper[4818]: E1122 04:47:52.496266 4818 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 04:47:52 crc kubenswrapper[4818]: E1122 04:47:52.597499 4818 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.625811 4818 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.699503 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.699549 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.699566 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.699588 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.699604 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:52Z","lastTransitionTime":"2025-11-22T04:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.732890 4818 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-11-22 04:42:51 +0000 UTC, rotation deadline is 2026-10-09 11:47:37.506103902 +0000 UTC Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.732942 4818 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7710h59m44.773165577s for next certificate rotation Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.802561 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.802618 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.802638 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.802663 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.802682 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:52Z","lastTransitionTime":"2025-11-22T04:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.904807 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.904847 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.904857 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.904874 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:52 crc kubenswrapper[4818]: I1122 04:47:52.904885 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:52Z","lastTransitionTime":"2025-11-22T04:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.007485 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.007543 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.007560 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.007583 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.007599 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:53Z","lastTransitionTime":"2025-11-22T04:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.110803 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.110844 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.110857 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.110872 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.110883 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:53Z","lastTransitionTime":"2025-11-22T04:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.185470 4818 apiserver.go:52] "Watching apiserver" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.203393 4818 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.203789 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-5rjvw","openshift-multus/multus-x865d","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-ovn-kubernetes/ovnkube-node-p8s7r","openshift-dns/node-resolver-ct4xj","openshift-multus/multus-additional-cni-plugins-j9r6v","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.204149 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.204217 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.204288 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.204343 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.204415 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.204652 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.205443 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.205586 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.205651 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.205906 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.206134 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.206212 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-ct4xj" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.205938 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.206870 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.208452 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.208701 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.209161 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.209349 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.209518 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.210246 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.210336 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.210890 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.210284 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.213959 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.214066 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.214184 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.214752 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.218718 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.218769 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.218802 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.218827 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.218835 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.218848 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.218857 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:53Z","lastTransitionTime":"2025-11-22T04:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.218928 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.218966 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.218997 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.219042 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.219075 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.219101 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.219127 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.219177 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.219224 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.219232 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.219266 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.219357 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.219488 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.219657 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.219694 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.219788 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.240245 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.249055 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.262245 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.274881 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.289631 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.298350 4818 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.303383 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.315660 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.321288 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.321501 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.321652 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.321800 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.321943 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:53Z","lastTransitionTime":"2025-11-22T04:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.326489 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.337912 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.350655 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.365337 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.386464 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.386522 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.386562 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.386592 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.386621 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.386649 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.386676 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.386706 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.386739 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.386768 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.386834 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.386860 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.386885 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.386914 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.386963 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387028 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387056 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387093 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387144 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387173 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387202 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387229 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387278 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387308 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387335 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387364 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387391 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387419 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387446 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387473 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387504 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387531 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387559 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387608 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387637 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387666 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387695 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387721 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387768 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387795 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387826 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387856 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387883 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387913 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387963 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387991 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388020 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388047 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388073 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388100 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388127 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388157 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388186 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388213 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388239 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388289 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388318 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388348 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388374 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388400 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388429 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388456 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388483 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388511 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388540 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388569 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388595 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388625 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388656 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388684 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388714 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388741 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388769 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388798 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388825 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388851 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388899 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388929 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388957 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388987 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389016 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389044 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389070 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389201 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389239 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389288 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389319 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389349 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389379 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389410 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389439 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389470 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389499 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389554 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389582 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389610 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389637 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389667 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389699 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389736 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389767 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389794 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389822 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389854 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389885 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389918 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389950 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389981 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390012 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390041 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390069 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390097 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390127 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390157 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390187 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390217 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390247 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390301 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390330 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390361 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390393 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390424 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390454 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390483 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390512 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390542 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390570 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390598 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390626 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390657 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390686 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390714 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390743 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390773 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390803 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390860 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390897 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390930 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390961 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.390991 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391022 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391051 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391082 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391116 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391144 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391178 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391210 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391245 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391314 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391345 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391372 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391403 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391436 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391467 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391496 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391527 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391560 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391590 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391625 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391656 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391687 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391720 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391753 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391786 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391820 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391853 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391887 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391920 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391951 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391985 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392019 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392050 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392080 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392112 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392147 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392182 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392219 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392278 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392311 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392347 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392380 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392428 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392460 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392493 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392527 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392560 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392593 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392626 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392661 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392696 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392729 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392762 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392793 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392827 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392864 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392966 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.393005 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.393078 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.393118 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.393152 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0f281330-68fd-4866-a787-83440a58aa4b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.393185 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-run-systemd\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.393215 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-node-log\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.393247 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0f281330-68fd-4866-a787-83440a58aa4b-cnibin\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.393304 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-host-var-lib-cni-bin\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.393334 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-host-var-lib-cni-multus\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.393365 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-hostroot\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.393405 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-multus-conf-dir\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.400366 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387037 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387437 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.403146 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.400951 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.387752 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.403351 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.388820 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389134 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389441 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.389994 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.391101 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392083 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.392652 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.393028 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.393204 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.393297 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.393980 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.394028 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.394142 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.399405 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.399685 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.399710 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.399861 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.400281 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.400297 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.400350 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.400355 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.400500 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.400660 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.400727 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.400847 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.400864 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.403361 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.403593 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.403450 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.403703 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.403899 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.403903 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.403999 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404023 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404058 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-slash\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404096 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404115 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0f281330-68fd-4866-a787-83440a58aa4b-system-cni-dir\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404171 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404203 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-run-ovn-kubernetes\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404240 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404289 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0f281330-68fd-4866-a787-83440a58aa4b-cni-binary-copy\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404321 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-cni-binary-copy\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404349 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/503d7b0b-1218-4c3b-b746-baea82235e2d-hosts-file\") pod \"node-resolver-ct4xj\" (UID: \"503d7b0b-1218-4c3b-b746-baea82235e2d\") " pod="openshift-dns/node-resolver-ct4xj" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404380 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404405 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0f281330-68fd-4866-a787-83440a58aa4b-os-release\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404432 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-host-run-k8s-cni-cncf-io\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404458 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-etc-kubernetes\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404485 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-var-lib-openvswitch\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404514 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-ovnkube-script-lib\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404540 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-multus-cni-dir\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404568 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b95fx\" (UniqueName: \"kubernetes.io/projected/0971c9ec-29ce-44eb-9875-ad4664143305-kube-api-access-b95fx\") pod \"machine-config-daemon-5rjvw\" (UID: \"0971c9ec-29ce-44eb-9875-ad4664143305\") " pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404597 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97n69\" (UniqueName: \"kubernetes.io/projected/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-kube-api-access-97n69\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404625 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89htt\" (UniqueName: \"kubernetes.io/projected/503d7b0b-1218-4c3b-b746-baea82235e2d-kube-api-access-89htt\") pod \"node-resolver-ct4xj\" (UID: \"503d7b0b-1218-4c3b-b746-baea82235e2d\") " pod="openshift-dns/node-resolver-ct4xj" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404656 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404685 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404716 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-host-run-netns\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404746 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-cni-bin\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404773 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-host-var-lib-kubelet\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404799 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-log-socket\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404828 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0971c9ec-29ce-44eb-9875-ad4664143305-proxy-tls\") pod \"machine-config-daemon-5rjvw\" (UID: \"0971c9ec-29ce-44eb-9875-ad4664143305\") " pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404853 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-cnibin\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404899 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-multus-socket-dir-parent\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404923 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-kubelet\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404949 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-run-openvswitch\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404978 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv2fh\" (UniqueName: \"kubernetes.io/projected/0f281330-68fd-4866-a787-83440a58aa4b-kube-api-access-hv2fh\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.404993 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.405014 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-run-netns\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.405122 4818 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.405763 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.406008 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.406001 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.406276 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.406501 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.406570 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.406713 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.406851 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.406871 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.406929 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.406970 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.406980 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.407070 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.407098 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.407145 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.407244 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.407394 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.407501 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.407523 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 04:47:53.907442761 +0000 UTC m=+26.481859328 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.407532 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.407708 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.407881 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.407930 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.408102 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.408122 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.408419 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.408684 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.408799 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.408794 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.409618 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.409840 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.410145 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.410144 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.410547 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.410551 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.410664 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.410731 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.410767 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.411002 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.411187 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.411221 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.411244 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.403821 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.419342 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.419754 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.420011 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.420701 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.421022 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.421200 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.421827 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.422066 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.422351 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.422690 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.423049 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.423087 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.423158 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.423389 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.425077 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.425377 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.425412 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.425480 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.425527 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.425906 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.426229 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.426273 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.426779 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.427165 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.427878 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.427957 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-system-cni-dir\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.428030 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-run-ovn\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.428101 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-os-release\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.428162 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0971c9ec-29ce-44eb-9875-ad4664143305-rootfs\") pod \"machine-config-daemon-5rjvw\" (UID: \"0971c9ec-29ce-44eb-9875-ad4664143305\") " pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.428225 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwd55\" (UniqueName: \"kubernetes.io/projected/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-kube-api-access-xwd55\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.428272 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-systemd-units\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.428318 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-cni-netd\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.428348 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0f281330-68fd-4866-a787-83440a58aa4b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.428998 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-multus-daemon-config\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.429056 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-host-run-multus-certs\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.429083 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-env-overrides\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.429115 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0971c9ec-29ce-44eb-9875-ad4664143305-mcd-auth-proxy-config\") pod \"machine-config-daemon-5rjvw\" (UID: \"0971c9ec-29ce-44eb-9875-ad4664143305\") " pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.429158 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-ovnkube-config\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.429194 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.430550 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.430595 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.430652 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.430742 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.430811 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-etc-openvswitch\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.430845 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-ovn-node-metrics-cert\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.430897 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.430932 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.430969 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.431185 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.431378 4818 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.431476 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 04:47:53.931457085 +0000 UTC m=+26.505873612 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.431674 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.431736 4818 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.431762 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.431782 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.431806 4818 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.431823 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.431839 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.431855 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.431876 4818 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.431897 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.431917 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.431937 4818 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.431953 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.431970 4818 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.431986 4818 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.432005 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.432020 4818 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.432035 4818 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.432049 4818 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.432067 4818 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.432083 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.432547 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.432620 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:47:53.932602055 +0000 UTC m=+26.507018582 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.433487 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.434345 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.435011 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.435217 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.435482 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.436186 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.436457 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.436711 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.436712 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.436919 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.437116 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.438052 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.438495 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.438634 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.438741 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.438844 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.438954 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:53Z","lastTransitionTime":"2025-11-22T04:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.440089 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.440484 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.440732 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.440824 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.440905 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.441175 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.403987 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.442554 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.442943 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.442980 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.443195 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.443277 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.445716 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.445752 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.445818 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.446701 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.447025 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.447240 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.447334 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.447472 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.447517 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.447867 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.448049 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.448277 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.448289 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.448372 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.448659 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.449939 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.450476 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.450651 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.451982 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.452071 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.452476 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.452668 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.452673 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.453109 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.453308 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.457986 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.463785 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.463984 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.464131 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.464360 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.464528 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.464618 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.464708 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.465227 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.466623 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.464481 4818 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.466854 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.466883 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.467119 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.467198 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.467242 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.467292 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.467318 4818 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.465595 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.467427 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 04:47:53.967369257 +0000 UTC m=+26.541785784 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.467606 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.468486 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.468686 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.470088 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.474746 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.476625 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.476634 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.478690 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.478717 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.478732 4818 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.479001 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 04:47:53.978798605 +0000 UTC m=+26.553215152 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.480552 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.481441 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.481500 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.481812 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.483617 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.484455 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.484454 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.484564 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.484622 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.485153 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.485286 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.485443 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.485807 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.486714 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.491398 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.491721 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.495456 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.507029 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.529129 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.532554 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.532589 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0971c9ec-29ce-44eb-9875-ad4664143305-mcd-auth-proxy-config\") pod \"machine-config-daemon-5rjvw\" (UID: \"0971c9ec-29ce-44eb-9875-ad4664143305\") " pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.532615 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-ovnkube-config\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.532637 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.532661 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-etc-openvswitch\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.532682 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-ovn-node-metrics-cert\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.532716 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.532729 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.532765 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.532799 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0f281330-68fd-4866-a787-83440a58aa4b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.532829 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-run-systemd\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.532905 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-node-log\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.532931 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0f281330-68fd-4866-a787-83440a58aa4b-system-cni-dir\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.532980 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0f281330-68fd-4866-a787-83440a58aa4b-cnibin\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.533056 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-host-var-lib-cni-bin\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.533085 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-host-var-lib-cni-multus\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.533108 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-hostroot\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.533155 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-multus-conf-dir\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.533177 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-slash\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.533232 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-run-ovn-kubernetes\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.533283 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0f281330-68fd-4866-a787-83440a58aa4b-os-release\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.533308 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0f281330-68fd-4866-a787-83440a58aa4b-cni-binary-copy\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.533329 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-cni-binary-copy\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.533377 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/503d7b0b-1218-4c3b-b746-baea82235e2d-hosts-file\") pod \"node-resolver-ct4xj\" (UID: \"503d7b0b-1218-4c3b-b746-baea82235e2d\") " pod="openshift-dns/node-resolver-ct4xj" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.533407 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-multus-cni-dir\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.533456 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-host-run-k8s-cni-cncf-io\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.533479 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-etc-kubernetes\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.533499 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-var-lib-openvswitch\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.533543 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0971c9ec-29ce-44eb-9875-ad4664143305-mcd-auth-proxy-config\") pod \"machine-config-daemon-5rjvw\" (UID: \"0971c9ec-29ce-44eb-9875-ad4664143305\") " pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.534231 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-ovnkube-script-lib\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.534300 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0f281330-68fd-4866-a787-83440a58aa4b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.534373 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-run-systemd\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.534432 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-node-log\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.534467 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0f281330-68fd-4866-a787-83440a58aa4b-system-cni-dir\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.534530 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0f281330-68fd-4866-a787-83440a58aa4b-cnibin\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.534614 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-host-var-lib-cni-bin\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.534669 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-host-var-lib-cni-multus\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.534707 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-hostroot\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.534736 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-multus-conf-dir\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.534794 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-slash\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.534799 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-ovnkube-config\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.534826 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-run-ovn-kubernetes\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.534856 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.534882 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0f281330-68fd-4866-a787-83440a58aa4b-os-release\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.534886 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-etc-openvswitch\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.533546 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-ovnkube-script-lib\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535351 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b95fx\" (UniqueName: \"kubernetes.io/projected/0971c9ec-29ce-44eb-9875-ad4664143305-kube-api-access-b95fx\") pod \"machine-config-daemon-5rjvw\" (UID: \"0971c9ec-29ce-44eb-9875-ad4664143305\") " pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535376 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97n69\" (UniqueName: \"kubernetes.io/projected/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-kube-api-access-97n69\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535419 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89htt\" (UniqueName: \"kubernetes.io/projected/503d7b0b-1218-4c3b-b746-baea82235e2d-kube-api-access-89htt\") pod \"node-resolver-ct4xj\" (UID: \"503d7b0b-1218-4c3b-b746-baea82235e2d\") " pod="openshift-dns/node-resolver-ct4xj" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535448 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-host-run-netns\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535466 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-log-socket\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535502 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-cni-bin\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535519 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-host-var-lib-kubelet\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535533 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv2fh\" (UniqueName: \"kubernetes.io/projected/0f281330-68fd-4866-a787-83440a58aa4b-kube-api-access-hv2fh\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535549 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0971c9ec-29ce-44eb-9875-ad4664143305-proxy-tls\") pod \"machine-config-daemon-5rjvw\" (UID: \"0971c9ec-29ce-44eb-9875-ad4664143305\") " pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535588 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-cnibin\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535603 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-multus-socket-dir-parent\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535618 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-kubelet\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535631 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-run-openvswitch\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535668 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-system-cni-dir\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535689 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-run-netns\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535734 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-run-ovn\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535755 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0971c9ec-29ce-44eb-9875-ad4664143305-rootfs\") pod \"machine-config-daemon-5rjvw\" (UID: \"0971c9ec-29ce-44eb-9875-ad4664143305\") " pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535776 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-os-release\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535794 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-multus-daemon-config\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535816 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwd55\" (UniqueName: \"kubernetes.io/projected/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-kube-api-access-xwd55\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535831 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-systemd-units\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535843 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-cni-netd\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535857 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0f281330-68fd-4866-a787-83440a58aa4b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535871 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-env-overrides\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535885 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-host-run-multus-certs\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535929 4818 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535940 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535950 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535959 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535966 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535974 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535983 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535990 4818 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535999 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536008 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536016 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536024 4818 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536033 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536042 4818 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536050 4818 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536050 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-cni-binary-copy\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536058 4818 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536077 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-host-run-multus-certs\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536088 4818 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536101 4818 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536113 4818 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536125 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536137 4818 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536149 4818 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536162 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536175 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536238 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536269 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536286 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536298 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536309 4818 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536321 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536333 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536346 4818 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536358 4818 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536369 4818 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536380 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536392 4818 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536404 4818 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536415 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536426 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536437 4818 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536448 4818 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536460 4818 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536471 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536483 4818 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536495 4818 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536507 4818 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536518 4818 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536528 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536539 4818 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536556 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536567 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536574 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-host-run-netns\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536578 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536606 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536613 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/503d7b0b-1218-4c3b-b746-baea82235e2d-hosts-file\") pod \"node-resolver-ct4xj\" (UID: \"503d7b0b-1218-4c3b-b746-baea82235e2d\") " pod="openshift-dns/node-resolver-ct4xj" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536619 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.535449 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0f281330-68fd-4866-a787-83440a58aa4b-cni-binary-copy\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536636 4818 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536660 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-log-socket\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536664 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536682 4818 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536692 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536703 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536703 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-multus-cni-dir\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536713 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536723 4818 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536734 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536746 4818 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536752 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-etc-kubernetes\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536756 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536774 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-cni-bin\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536776 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536794 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536795 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-var-lib-openvswitch\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536805 4818 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536816 4818 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536825 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-run-netns\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536828 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536873 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536883 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-run-ovn\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536888 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536901 4818 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536913 4818 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536925 4818 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536937 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536948 4818 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536953 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-os-release\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536960 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536973 4818 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536986 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536998 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537010 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537022 4818 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537033 4818 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537046 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537059 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537071 4818 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537083 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537095 4818 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537107 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537118 4818 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537130 4818 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537142 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537154 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537166 4818 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537177 4818 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537188 4818 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537200 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537211 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537222 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537233 4818 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537245 4818 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537272 4818 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537285 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537295 4818 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537306 4818 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537317 4818 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537328 4818 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537340 4818 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537352 4818 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537364 4818 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537375 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537387 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537398 4818 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537409 4818 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537420 4818 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537431 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537441 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537452 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537462 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537473 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537484 4818 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537497 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537504 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-multus-daemon-config\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537508 4818 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537532 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-host-var-lib-kubelet\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537541 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537554 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537592 4818 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537603 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537614 4818 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537623 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537633 4818 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537670 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537684 4818 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537695 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537707 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537718 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537754 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537764 4818 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537774 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537783 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537793 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537801 4818 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537828 4818 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537840 4818 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537852 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537863 4818 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537873 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537884 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537922 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537935 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537949 4818 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537960 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.537970 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538010 4818 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538020 4818 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538031 4818 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538043 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538084 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538099 4818 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538114 4818 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538125 4818 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538137 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538179 4818 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538191 4818 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538201 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538212 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538221 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538267 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538280 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538291 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538301 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538337 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538350 4818 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538360 4818 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538371 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536912 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0971c9ec-29ce-44eb-9875-ad4664143305-rootfs\") pod \"machine-config-daemon-5rjvw\" (UID: \"0971c9ec-29ce-44eb-9875-ad4664143305\") " pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538632 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-systemd-units\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.538695 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-cni-netd\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.541139 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0f281330-68fd-4866-a787-83440a58aa4b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.541639 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-env-overrides\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.541697 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-kubelet\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.541729 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-cnibin\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.541780 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-multus-socket-dir-parent\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.541811 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-system-cni-dir\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.541830 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-run-openvswitch\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.536733 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-host-run-k8s-cni-cncf-io\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.542968 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.543085 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0971c9ec-29ce-44eb-9875-ad4664143305-proxy-tls\") pod \"machine-config-daemon-5rjvw\" (UID: \"0971c9ec-29ce-44eb-9875-ad4664143305\") " pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.543209 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.543226 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.543236 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.543267 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.543277 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:53Z","lastTransitionTime":"2025-11-22T04:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.548165 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-ovn-node-metrics-cert\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.552704 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97n69\" (UniqueName: \"kubernetes.io/projected/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-kube-api-access-97n69\") pod \"ovnkube-node-p8s7r\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.554591 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89htt\" (UniqueName: \"kubernetes.io/projected/503d7b0b-1218-4c3b-b746-baea82235e2d-kube-api-access-89htt\") pod \"node-resolver-ct4xj\" (UID: \"503d7b0b-1218-4c3b-b746-baea82235e2d\") " pod="openshift-dns/node-resolver-ct4xj" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.554998 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv2fh\" (UniqueName: \"kubernetes.io/projected/0f281330-68fd-4866-a787-83440a58aa4b-kube-api-access-hv2fh\") pod \"multus-additional-cni-plugins-j9r6v\" (UID: \"0f281330-68fd-4866-a787-83440a58aa4b\") " pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.557823 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b95fx\" (UniqueName: \"kubernetes.io/projected/0971c9ec-29ce-44eb-9875-ad4664143305-kube-api-access-b95fx\") pod \"machine-config-daemon-5rjvw\" (UID: \"0971c9ec-29ce-44eb-9875-ad4664143305\") " pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.559147 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwd55\" (UniqueName: \"kubernetes.io/projected/d96879f8-2766-4f5e-bc3e-bbf8e5394a94-kube-api-access-xwd55\") pod \"multus-x865d\" (UID: \"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\") " pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.570843 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.583948 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-ct4xj" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.600126 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-x865d" Nov 22 04:47:53 crc kubenswrapper[4818]: W1122 04:47:53.600553 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8f88a0c_c404_4e82_88c5_3e0f4f2b9025.slice/crio-0f7118fd73ae73c6226c848498ad1acd4cf0c3b128a39b62ab344e4193d7973c WatchSource:0}: Error finding container 0f7118fd73ae73c6226c848498ad1acd4cf0c3b128a39b62ab344e4193d7973c: Status 404 returned error can't find the container with id 0f7118fd73ae73c6226c848498ad1acd4cf0c3b128a39b62ab344e4193d7973c Nov 22 04:47:53 crc kubenswrapper[4818]: W1122 04:47:53.600956 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod503d7b0b_1218_4c3b_b746_baea82235e2d.slice/crio-71fc3edae04aa2356568410032b404df6064c69a81b163c643ba13109ead2eb0 WatchSource:0}: Error finding container 71fc3edae04aa2356568410032b404df6064c69a81b163c643ba13109ead2eb0: Status 404 returned error can't find the container with id 71fc3edae04aa2356568410032b404df6064c69a81b163c643ba13109ead2eb0 Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.613063 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" Nov 22 04:47:53 crc kubenswrapper[4818]: W1122 04:47:53.618524 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd96879f8_2766_4f5e_bc3e_bbf8e5394a94.slice/crio-4946aee289ac8df5dc6abed9fb551038272d4732cdba3730704701d8e91eef12 WatchSource:0}: Error finding container 4946aee289ac8df5dc6abed9fb551038272d4732cdba3730704701d8e91eef12: Status 404 returned error can't find the container with id 4946aee289ac8df5dc6abed9fb551038272d4732cdba3730704701d8e91eef12 Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.645978 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.646384 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.646396 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.646432 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.646443 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:53Z","lastTransitionTime":"2025-11-22T04:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.748945 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.748975 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.748983 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.748996 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.749005 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:53Z","lastTransitionTime":"2025-11-22T04:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.816380 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.851377 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.851421 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.851432 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.851451 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.851463 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:53Z","lastTransitionTime":"2025-11-22T04:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.856747 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 04:47:53 crc kubenswrapper[4818]: W1122 04:47:53.876842 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-2d40f755e86c46a0440c0c3dfc6b491fd47b1db2a0782eababc8374345f5b616 WatchSource:0}: Error finding container 2d40f755e86c46a0440c0c3dfc6b491fd47b1db2a0782eababc8374345f5b616: Status 404 returned error can't find the container with id 2d40f755e86c46a0440c0c3dfc6b491fd47b1db2a0782eababc8374345f5b616 Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.942848 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.942983 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:47:54.942961905 +0000 UTC m=+27.517378432 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.943051 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.943110 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.943224 4818 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.943282 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 04:47:54.943275493 +0000 UTC m=+27.517692020 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.943564 4818 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 04:47:53 crc kubenswrapper[4818]: E1122 04:47:53.943953 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 04:47:54.943873199 +0000 UTC m=+27.518289806 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.954526 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.954563 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.954575 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.954590 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:53 crc kubenswrapper[4818]: I1122 04:47:53.954601 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:53Z","lastTransitionTime":"2025-11-22T04:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.045330 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.045403 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:47:54 crc kubenswrapper[4818]: E1122 04:47:54.045555 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 04:47:54 crc kubenswrapper[4818]: E1122 04:47:54.045559 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 04:47:54 crc kubenswrapper[4818]: E1122 04:47:54.045569 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 04:47:54 crc kubenswrapper[4818]: E1122 04:47:54.045589 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 04:47:54 crc kubenswrapper[4818]: E1122 04:47:54.045600 4818 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:47:54 crc kubenswrapper[4818]: E1122 04:47:54.045605 4818 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:47:54 crc kubenswrapper[4818]: E1122 04:47:54.045648 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 04:47:55.045635253 +0000 UTC m=+27.620051780 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:47:54 crc kubenswrapper[4818]: E1122 04:47:54.045684 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 04:47:55.045676825 +0000 UTC m=+27.620093352 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.056777 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.056823 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.056833 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.056846 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.056854 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:54Z","lastTransitionTime":"2025-11-22T04:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.159859 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.159899 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.159909 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.159924 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.159935 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:54Z","lastTransitionTime":"2025-11-22T04:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.265908 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.265971 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.265983 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.265999 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.266009 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:54Z","lastTransitionTime":"2025-11-22T04:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.291046 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:47:54 crc kubenswrapper[4818]: E1122 04:47:54.291173 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.295177 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.295859 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.297006 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.297664 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.298636 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.299159 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.299720 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.300635 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.301501 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.302432 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.302930 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.304079 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.304600 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.305162 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.306026 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.306555 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.307555 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.308033 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.308625 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.309621 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.310070 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.310970 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.311475 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.312653 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.313172 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.313804 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.314882 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.315355 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.316239 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.316785 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.317698 4818 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.317817 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.319947 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.321032 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.321598 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.323151 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.323956 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.324887 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.325579 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.326769 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.327575 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.328547 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.329197 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.330186 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.330655 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.331552 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.332082 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.333409 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.333954 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.334948 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.335402 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.336330 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.336887 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.337338 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.338094 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-n4wst"] Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.338543 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-n4wst" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.340294 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.340531 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.341520 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.341930 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.350511 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.362051 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.367830 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.367861 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.367872 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.367887 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.367899 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:54Z","lastTransitionTime":"2025-11-22T04:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.374838 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.384243 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.393097 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.401187 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.408053 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.417224 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.427663 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.442173 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.449861 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/84fcc74a-fc3b-4885-9c82-6f3956d50c92-serviceca\") pod \"node-ca-n4wst\" (UID: \"84fcc74a-fc3b-4885-9c82-6f3956d50c92\") " pod="openshift-image-registry/node-ca-n4wst" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.449920 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff46s\" (UniqueName: \"kubernetes.io/projected/84fcc74a-fc3b-4885-9c82-6f3956d50c92-kube-api-access-ff46s\") pod \"node-ca-n4wst\" (UID: \"84fcc74a-fc3b-4885-9c82-6f3956d50c92\") " pod="openshift-image-registry/node-ca-n4wst" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.449943 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84fcc74a-fc3b-4885-9c82-6f3956d50c92-host\") pod \"node-ca-n4wst\" (UID: \"84fcc74a-fc3b-4885-9c82-6f3956d50c92\") " pod="openshift-image-registry/node-ca-n4wst" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.452599 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.459344 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.463837 4818 generic.go:334] "Generic (PLEG): container finished" podID="0f281330-68fd-4866-a787-83440a58aa4b" containerID="ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b" exitCode=0 Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.463921 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" event={"ID":"0f281330-68fd-4866-a787-83440a58aa4b","Type":"ContainerDied","Data":"ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.463994 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" event={"ID":"0f281330-68fd-4866-a787-83440a58aa4b","Type":"ContainerStarted","Data":"22ca8be9e19b490bf63438276aa770f565a1c7e6679dee4a947cdc8610775ba9"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.465371 4818 generic.go:334] "Generic (PLEG): container finished" podID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerID="6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb" exitCode=0 Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.465460 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerDied","Data":"6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.465505 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerStarted","Data":"0f7118fd73ae73c6226c848498ad1acd4cf0c3b128a39b62ab344e4193d7973c"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.467831 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"7aee7c4aab8182ebfec2218c48faee0fe5e8bdb45a459319efc0c0ad6c826436"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.472875 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.472929 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.472945 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.472967 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.472989 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:54Z","lastTransitionTime":"2025-11-22T04:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.474181 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.474227 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"2d40f755e86c46a0440c0c3dfc6b491fd47b1db2a0782eababc8374345f5b616"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.481370 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.481390 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.481457 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.481477 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"2565547222a3aa89cca131cc20beb06842dab4d625986c36d54a2e303d0ba762"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.486186 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x865d" event={"ID":"d96879f8-2766-4f5e-bc3e-bbf8e5394a94","Type":"ContainerStarted","Data":"27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.486213 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x865d" event={"ID":"d96879f8-2766-4f5e-bc3e-bbf8e5394a94","Type":"ContainerStarted","Data":"4946aee289ac8df5dc6abed9fb551038272d4732cdba3730704701d8e91eef12"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.488816 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-ct4xj" event={"ID":"503d7b0b-1218-4c3b-b746-baea82235e2d","Type":"ContainerStarted","Data":"90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.488844 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-ct4xj" event={"ID":"503d7b0b-1218-4c3b-b746-baea82235e2d","Type":"ContainerStarted","Data":"71fc3edae04aa2356568410032b404df6064c69a81b163c643ba13109ead2eb0"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.490788 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.490816 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.490829 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5ae110befebbae6052affe2b07bfc80529ad489bba5ee33e84b7592f6d5abd20"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.493267 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.538138 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.553782 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/84fcc74a-fc3b-4885-9c82-6f3956d50c92-serviceca\") pod \"node-ca-n4wst\" (UID: \"84fcc74a-fc3b-4885-9c82-6f3956d50c92\") " pod="openshift-image-registry/node-ca-n4wst" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.553840 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff46s\" (UniqueName: \"kubernetes.io/projected/84fcc74a-fc3b-4885-9c82-6f3956d50c92-kube-api-access-ff46s\") pod \"node-ca-n4wst\" (UID: \"84fcc74a-fc3b-4885-9c82-6f3956d50c92\") " pod="openshift-image-registry/node-ca-n4wst" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.553874 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84fcc74a-fc3b-4885-9c82-6f3956d50c92-host\") pod \"node-ca-n4wst\" (UID: \"84fcc74a-fc3b-4885-9c82-6f3956d50c92\") " pod="openshift-image-registry/node-ca-n4wst" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.555587 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84fcc74a-fc3b-4885-9c82-6f3956d50c92-host\") pod \"node-ca-n4wst\" (UID: \"84fcc74a-fc3b-4885-9c82-6f3956d50c92\") " pod="openshift-image-registry/node-ca-n4wst" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.556511 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/84fcc74a-fc3b-4885-9c82-6f3956d50c92-serviceca\") pod \"node-ca-n4wst\" (UID: \"84fcc74a-fc3b-4885-9c82-6f3956d50c92\") " pod="openshift-image-registry/node-ca-n4wst" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.563588 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.574185 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.576416 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.576879 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.577018 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.577136 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.577231 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:54Z","lastTransitionTime":"2025-11-22T04:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.578925 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff46s\" (UniqueName: \"kubernetes.io/projected/84fcc74a-fc3b-4885-9c82-6f3956d50c92-kube-api-access-ff46s\") pod \"node-ca-n4wst\" (UID: \"84fcc74a-fc3b-4885-9c82-6f3956d50c92\") " pod="openshift-image-registry/node-ca-n4wst" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.607327 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.626096 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.637971 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.645851 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.652677 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-n4wst" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.656143 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.678232 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.679705 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.679729 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.679737 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.679750 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.679758 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:54Z","lastTransitionTime":"2025-11-22T04:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.685897 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.693093 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.703207 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.715175 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.727174 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.738767 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.749070 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.767672 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:54Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.783035 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.783074 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.783083 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.783098 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.783107 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:54Z","lastTransitionTime":"2025-11-22T04:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.786573 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:54Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.806644 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:54Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.822904 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:54Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.831353 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:54Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.843854 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:54Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.884851 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.884887 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.884899 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.884915 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.884927 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:54Z","lastTransitionTime":"2025-11-22T04:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.956851 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.956971 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:47:54 crc kubenswrapper[4818]: E1122 04:47:54.957006 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:47:56.956984783 +0000 UTC m=+29.531401320 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.957067 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:47:54 crc kubenswrapper[4818]: E1122 04:47:54.957072 4818 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 04:47:54 crc kubenswrapper[4818]: E1122 04:47:54.957140 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 04:47:56.957130756 +0000 UTC m=+29.531547283 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 04:47:54 crc kubenswrapper[4818]: E1122 04:47:54.957174 4818 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 04:47:54 crc kubenswrapper[4818]: E1122 04:47:54.957224 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 04:47:56.957207909 +0000 UTC m=+29.531624466 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.987645 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.987679 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.987688 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.987702 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:54 crc kubenswrapper[4818]: I1122 04:47:54.987711 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:54Z","lastTransitionTime":"2025-11-22T04:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.058592 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.058671 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:47:55 crc kubenswrapper[4818]: E1122 04:47:55.058796 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 04:47:55 crc kubenswrapper[4818]: E1122 04:47:55.058824 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 04:47:55 crc kubenswrapper[4818]: E1122 04:47:55.058836 4818 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:47:55 crc kubenswrapper[4818]: E1122 04:47:55.058838 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 04:47:55 crc kubenswrapper[4818]: E1122 04:47:55.058868 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 04:47:55 crc kubenswrapper[4818]: E1122 04:47:55.058885 4818 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:47:55 crc kubenswrapper[4818]: E1122 04:47:55.058891 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 04:47:57.05887632 +0000 UTC m=+29.633292847 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:47:55 crc kubenswrapper[4818]: E1122 04:47:55.058949 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 04:47:57.058928362 +0000 UTC m=+29.633344969 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.089970 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.090005 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.090014 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.090027 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.090036 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:55Z","lastTransitionTime":"2025-11-22T04:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.191871 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.191919 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.191932 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.191951 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.191964 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:55Z","lastTransitionTime":"2025-11-22T04:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.290968 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.290982 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:47:55 crc kubenswrapper[4818]: E1122 04:47:55.291130 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:47:55 crc kubenswrapper[4818]: E1122 04:47:55.291193 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.294593 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.294622 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.294630 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.294642 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.294652 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:55Z","lastTransitionTime":"2025-11-22T04:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.396667 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.396710 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.396721 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.396737 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.396749 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:55Z","lastTransitionTime":"2025-11-22T04:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.498202 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.498242 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.498279 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.498313 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.498328 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:55Z","lastTransitionTime":"2025-11-22T04:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.500074 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerStarted","Data":"7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761"} Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.500124 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerStarted","Data":"051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b"} Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.500139 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerStarted","Data":"f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7"} Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.500152 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerStarted","Data":"c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397"} Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.500165 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerStarted","Data":"d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc"} Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.502077 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-n4wst" event={"ID":"84fcc74a-fc3b-4885-9c82-6f3956d50c92","Type":"ContainerStarted","Data":"fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63"} Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.502117 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-n4wst" event={"ID":"84fcc74a-fc3b-4885-9c82-6f3956d50c92","Type":"ContainerStarted","Data":"fef815e2ddbb3e07662c01af1ca8a31caf2b9c431e6a9d55286755ab0d3a67ef"} Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.509526 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" event={"ID":"0f281330-68fd-4866-a787-83440a58aa4b","Type":"ContainerStarted","Data":"8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5"} Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.525819 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.549863 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.566605 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.579201 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.592416 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.600580 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.600635 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.600651 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.600671 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.600686 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:55Z","lastTransitionTime":"2025-11-22T04:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.609941 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.623976 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.634981 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.651501 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.665027 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.683407 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.702244 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.703140 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.703183 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.703196 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.703212 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.703225 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:55Z","lastTransitionTime":"2025-11-22T04:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.718923 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.732651 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.750000 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.768031 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.783139 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.798672 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.806202 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.806296 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.806315 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.806343 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.806361 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:55Z","lastTransitionTime":"2025-11-22T04:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.820023 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.838664 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.865013 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.879477 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.894146 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.909474 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.909530 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.909549 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.909570 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.909586 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:55Z","lastTransitionTime":"2025-11-22T04:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:55 crc kubenswrapper[4818]: I1122 04:47:55.911664 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:55Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.012665 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.012754 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.012772 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.012793 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.012804 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:56Z","lastTransitionTime":"2025-11-22T04:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.116515 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.116580 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.116602 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.116632 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.116655 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:56Z","lastTransitionTime":"2025-11-22T04:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.220100 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.220151 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.220165 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.220184 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.220198 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:56Z","lastTransitionTime":"2025-11-22T04:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.292495 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:47:56 crc kubenswrapper[4818]: E1122 04:47:56.292652 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.323058 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.323097 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.323110 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.323126 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.323139 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:56Z","lastTransitionTime":"2025-11-22T04:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.425686 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.425727 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.425740 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.425756 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.425770 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:56Z","lastTransitionTime":"2025-11-22T04:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.517657 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerStarted","Data":"8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e"} Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.528837 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.528884 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.528901 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.528921 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.528937 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:56Z","lastTransitionTime":"2025-11-22T04:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.631490 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.631553 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.631570 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.631593 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.631609 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:56Z","lastTransitionTime":"2025-11-22T04:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.734662 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.735103 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.735123 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.735146 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.735166 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:56Z","lastTransitionTime":"2025-11-22T04:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.838446 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.838506 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.838525 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.838555 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.838578 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:56Z","lastTransitionTime":"2025-11-22T04:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.941840 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.941878 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.941888 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.941903 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.941914 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:56Z","lastTransitionTime":"2025-11-22T04:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.980786 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.980947 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:47:56 crc kubenswrapper[4818]: E1122 04:47:56.981024 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:48:00.980989003 +0000 UTC m=+33.555405540 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:47:56 crc kubenswrapper[4818]: E1122 04:47:56.981070 4818 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 04:47:56 crc kubenswrapper[4818]: I1122 04:47:56.981124 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:47:56 crc kubenswrapper[4818]: E1122 04:47:56.981136 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 04:48:00.981116406 +0000 UTC m=+33.555533023 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 04:47:56 crc kubenswrapper[4818]: E1122 04:47:56.981295 4818 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 04:47:56 crc kubenswrapper[4818]: E1122 04:47:56.981351 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 04:48:00.981340892 +0000 UTC m=+33.555757519 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.044056 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.044134 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.044187 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.044213 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.044236 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:57Z","lastTransitionTime":"2025-11-22T04:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.082145 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.082226 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:47:57 crc kubenswrapper[4818]: E1122 04:47:57.082406 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 04:47:57 crc kubenswrapper[4818]: E1122 04:47:57.082407 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 04:47:57 crc kubenswrapper[4818]: E1122 04:47:57.082431 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 04:47:57 crc kubenswrapper[4818]: E1122 04:47:57.082446 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 04:47:57 crc kubenswrapper[4818]: E1122 04:47:57.082455 4818 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:47:57 crc kubenswrapper[4818]: E1122 04:47:57.082465 4818 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:47:57 crc kubenswrapper[4818]: E1122 04:47:57.082529 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 04:48:01.08250763 +0000 UTC m=+33.656924197 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:47:57 crc kubenswrapper[4818]: E1122 04:47:57.082555 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 04:48:01.082543521 +0000 UTC m=+33.656960088 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.146793 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.146851 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.146868 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.146889 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.146907 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:57Z","lastTransitionTime":"2025-11-22T04:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.249408 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.249449 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.249461 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.249477 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.249487 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:57Z","lastTransitionTime":"2025-11-22T04:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.290493 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.290591 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:47:57 crc kubenswrapper[4818]: E1122 04:47:57.290714 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:47:57 crc kubenswrapper[4818]: E1122 04:47:57.290880 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.351487 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.351527 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.351539 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.351558 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.351569 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:57Z","lastTransitionTime":"2025-11-22T04:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.454584 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.454828 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.454836 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.454850 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.454861 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:57Z","lastTransitionTime":"2025-11-22T04:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.522797 4818 generic.go:334] "Generic (PLEG): container finished" podID="0f281330-68fd-4866-a787-83440a58aa4b" containerID="8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5" exitCode=0 Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.522855 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" event={"ID":"0f281330-68fd-4866-a787-83440a58aa4b","Type":"ContainerDied","Data":"8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5"} Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.524160 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1"} Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.537234 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.556041 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.557548 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.557585 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.557598 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.557696 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.557722 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:57Z","lastTransitionTime":"2025-11-22T04:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.569233 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.582102 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.599759 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.615558 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.625364 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.633762 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.646501 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.657321 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.659951 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.659981 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.659993 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.660009 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.660022 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:57Z","lastTransitionTime":"2025-11-22T04:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.675403 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.698206 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.728837 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.755138 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.763351 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.763596 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.763667 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.763735 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.763884 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:57Z","lastTransitionTime":"2025-11-22T04:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.779199 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.789686 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.802163 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.810576 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.821871 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.835211 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.853529 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.865781 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.865818 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.865830 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.865846 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.865854 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:57Z","lastTransitionTime":"2025-11-22T04:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.869663 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.885726 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.902722 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:57Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.968622 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.968658 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.968668 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.968683 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:57 crc kubenswrapper[4818]: I1122 04:47:57.968695 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:57Z","lastTransitionTime":"2025-11-22T04:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.015442 4818 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.077283 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.077338 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.077354 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.077378 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.077395 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:58Z","lastTransitionTime":"2025-11-22T04:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.179067 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.179126 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.179143 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.179168 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.179186 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:58Z","lastTransitionTime":"2025-11-22T04:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.282334 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.282392 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.282410 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.282432 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.282449 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:58Z","lastTransitionTime":"2025-11-22T04:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.290725 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:47:58 crc kubenswrapper[4818]: E1122 04:47:58.290818 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.311716 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.332597 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.348147 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.369600 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.384716 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.384770 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.384781 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.384794 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.384807 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:58Z","lastTransitionTime":"2025-11-22T04:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.388691 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.400184 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.413580 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.425273 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.443428 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.464497 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.478530 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.487272 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.487296 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.487304 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.487316 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.487326 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:58Z","lastTransitionTime":"2025-11-22T04:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.492141 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.530050 4818 generic.go:334] "Generic (PLEG): container finished" podID="0f281330-68fd-4866-a787-83440a58aa4b" containerID="89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08" exitCode=0 Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.530227 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" event={"ID":"0f281330-68fd-4866-a787-83440a58aa4b","Type":"ContainerDied","Data":"89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08"} Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.538980 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerStarted","Data":"08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00"} Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.551260 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.569747 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.580778 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.590063 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.590109 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.590118 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.590132 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.590142 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:58Z","lastTransitionTime":"2025-11-22T04:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.594997 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.607347 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.610096 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.617941 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.624685 4818 scope.go:117] "RemoveContainer" containerID="e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe" Nov 22 04:47:58 crc kubenswrapper[4818]: E1122 04:47:58.624983 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.627730 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.632714 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.645906 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.673467 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.690799 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.692407 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.692483 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.692499 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.692515 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.692526 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:58Z","lastTransitionTime":"2025-11-22T04:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.702903 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.716639 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.725748 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.739105 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.759270 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.774336 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.787531 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.795441 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.795486 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.795498 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.795517 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.795530 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:58Z","lastTransitionTime":"2025-11-22T04:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.800875 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.812394 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.823950 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.833642 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.845535 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.860390 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.889035 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.898429 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.898476 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.898489 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.898510 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.898523 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:58Z","lastTransitionTime":"2025-11-22T04:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:58 crc kubenswrapper[4818]: I1122 04:47:58.906399 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.001691 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.001745 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.001765 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.001789 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.001806 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:59Z","lastTransitionTime":"2025-11-22T04:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.105045 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.105100 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.105116 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.105141 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.105158 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:59Z","lastTransitionTime":"2025-11-22T04:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.208530 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.208651 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.208678 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.208705 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.208725 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:59Z","lastTransitionTime":"2025-11-22T04:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.290445 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.290509 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:47:59 crc kubenswrapper[4818]: E1122 04:47:59.290610 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:47:59 crc kubenswrapper[4818]: E1122 04:47:59.290712 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.311798 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.311846 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.311868 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.311890 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.311909 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:59Z","lastTransitionTime":"2025-11-22T04:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.415693 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.415759 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.415785 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.415812 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.415833 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:59Z","lastTransitionTime":"2025-11-22T04:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.518508 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.518556 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.518574 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.518594 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.518610 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:59Z","lastTransitionTime":"2025-11-22T04:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.551600 4818 generic.go:334] "Generic (PLEG): container finished" podID="0f281330-68fd-4866-a787-83440a58aa4b" containerID="0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a" exitCode=0 Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.551683 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" event={"ID":"0f281330-68fd-4866-a787-83440a58aa4b","Type":"ContainerDied","Data":"0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a"} Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.552727 4818 scope.go:117] "RemoveContainer" containerID="e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe" Nov 22 04:47:59 crc kubenswrapper[4818]: E1122 04:47:59.553012 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.578369 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:59Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.602613 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:59Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.621806 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.621883 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.621908 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.621943 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.621970 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:59Z","lastTransitionTime":"2025-11-22T04:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.622226 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:59Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.636808 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:59Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.655259 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:59Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.682859 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:59Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.697150 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:59Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.711122 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:59Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.724438 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:59Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.724498 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.724537 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.724548 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.724565 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.724576 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:59Z","lastTransitionTime":"2025-11-22T04:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.743779 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:59Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.765521 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:59Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.781172 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:59Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.792873 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:47:59Z is after 2025-08-24T17:21:41Z" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.826316 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.826357 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.826368 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.826383 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.826394 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:59Z","lastTransitionTime":"2025-11-22T04:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.929450 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.929488 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.929499 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.929515 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:47:59 crc kubenswrapper[4818]: I1122 04:47:59.929529 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:47:59Z","lastTransitionTime":"2025-11-22T04:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.032782 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.032816 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.032824 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.032837 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.032846 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:00Z","lastTransitionTime":"2025-11-22T04:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.136475 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.136517 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.136528 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.136543 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.136556 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:00Z","lastTransitionTime":"2025-11-22T04:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.238732 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.238767 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.238776 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.238789 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.238797 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:00Z","lastTransitionTime":"2025-11-22T04:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.293895 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:00 crc kubenswrapper[4818]: E1122 04:48:00.294352 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.341216 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.341279 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.341295 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.341311 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.341323 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:00Z","lastTransitionTime":"2025-11-22T04:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.443181 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.443213 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.443221 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.443235 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.443247 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:00Z","lastTransitionTime":"2025-11-22T04:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.545911 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.545933 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.545941 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.545953 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.545963 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:00Z","lastTransitionTime":"2025-11-22T04:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.558027 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" event={"ID":"0f281330-68fd-4866-a787-83440a58aa4b","Type":"ContainerStarted","Data":"436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4"} Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.564746 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerStarted","Data":"120854347bb61d6824e445e5af011b9a91988e9334cd9b0855c9e8a9282028cb"} Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.565240 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.575605 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.593970 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.595769 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.615794 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.634123 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.649483 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.649965 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.649993 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.650005 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.650022 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.650040 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:00Z","lastTransitionTime":"2025-11-22T04:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.667037 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.687147 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.707230 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.722197 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.734668 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.747024 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.752824 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.752864 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.752879 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.752896 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.752908 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:00Z","lastTransitionTime":"2025-11-22T04:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.766041 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.777319 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.793066 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.806470 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.820202 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.839302 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.852068 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.856725 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.856812 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.856842 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.856930 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.856948 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:00Z","lastTransitionTime":"2025-11-22T04:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.867452 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.881049 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.892665 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.912818 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.927232 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.954878 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://120854347bb61d6824e445e5af011b9a91988e9334cd9b0855c9e8a9282028cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.958726 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.958756 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.958768 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.958782 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.958793 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:00Z","lastTransitionTime":"2025-11-22T04:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.976904 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:00 crc kubenswrapper[4818]: I1122 04:48:00.994149 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:00Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.035059 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.035171 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.035228 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:48:09.035205625 +0000 UTC m=+41.609622152 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.035241 4818 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.035346 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 04:48:09.035330088 +0000 UTC m=+41.609746625 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.035383 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.035481 4818 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.035524 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 04:48:09.035517003 +0000 UTC m=+41.609933530 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.061068 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.061100 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.061109 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.061121 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.061130 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:01Z","lastTransitionTime":"2025-11-22T04:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.136143 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.136227 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.136431 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.136463 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.136459 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.136541 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.136571 4818 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.136665 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 04:48:09.13663502 +0000 UTC m=+41.711051577 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.136484 4818 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.136771 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 04:48:09.136752573 +0000 UTC m=+41.711169230 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.163642 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.163678 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.163687 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.163701 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.163709 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:01Z","lastTransitionTime":"2025-11-22T04:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.266434 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.266508 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.266526 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.266549 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.266565 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:01Z","lastTransitionTime":"2025-11-22T04:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.291076 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.291086 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.291372 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.291212 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.294210 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.294280 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.294301 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.294320 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.294332 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:01Z","lastTransitionTime":"2025-11-22T04:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.308549 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.312824 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.312881 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.312894 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.312915 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.312930 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:01Z","lastTransitionTime":"2025-11-22T04:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.327011 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.331232 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.331358 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.331387 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.331420 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.331443 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:01Z","lastTransitionTime":"2025-11-22T04:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.346750 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.351540 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.351595 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.351614 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.351635 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.351652 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:01Z","lastTransitionTime":"2025-11-22T04:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.370098 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.403357 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.403396 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.403409 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.403423 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.403434 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:01Z","lastTransitionTime":"2025-11-22T04:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.427525 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: E1122 04:48:01.427646 4818 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.429590 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.429614 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.429622 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.429635 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.429644 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:01Z","lastTransitionTime":"2025-11-22T04:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.532052 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.532095 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.532107 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.532124 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.532136 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:01Z","lastTransitionTime":"2025-11-22T04:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.576279 4818 generic.go:334] "Generic (PLEG): container finished" podID="0f281330-68fd-4866-a787-83440a58aa4b" containerID="436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4" exitCode=0 Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.576340 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" event={"ID":"0f281330-68fd-4866-a787-83440a58aa4b","Type":"ContainerDied","Data":"436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4"} Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.576484 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.577240 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.594107 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.609066 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.616673 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://120854347bb61d6824e445e5af011b9a91988e9334cd9b0855c9e8a9282028cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.632892 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.634774 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.634805 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.634817 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.634843 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.634855 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:01Z","lastTransitionTime":"2025-11-22T04:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.645667 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.657339 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.674275 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.686767 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.700828 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.710326 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.721394 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.730546 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.739428 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.739563 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.739633 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.739697 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.739758 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:01Z","lastTransitionTime":"2025-11-22T04:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.741142 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.751507 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.762995 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.774636 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.785961 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.796805 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.807075 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.818894 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.834615 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.842461 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.842529 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.842547 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.842570 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.842590 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:01Z","lastTransitionTime":"2025-11-22T04:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.847291 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.871699 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://120854347bb61d6824e445e5af011b9a91988e9334cd9b0855c9e8a9282028cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.885841 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.896354 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.906982 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.915770 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:01Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.946391 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.946427 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.946444 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.946460 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:01 crc kubenswrapper[4818]: I1122 04:48:01.946470 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:01Z","lastTransitionTime":"2025-11-22T04:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.050032 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.050076 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.050087 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.050102 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.050113 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:02Z","lastTransitionTime":"2025-11-22T04:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.153301 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.153384 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.153407 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.153434 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.153451 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:02Z","lastTransitionTime":"2025-11-22T04:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.255468 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.255511 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.255522 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.255540 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.255551 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:02Z","lastTransitionTime":"2025-11-22T04:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.290822 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:02 crc kubenswrapper[4818]: E1122 04:48:02.291315 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.358681 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.358729 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.358746 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.358769 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.358785 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:02Z","lastTransitionTime":"2025-11-22T04:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.461859 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.461904 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.461921 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.461943 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.461959 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:02Z","lastTransitionTime":"2025-11-22T04:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.564634 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.564686 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.564703 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.564725 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.564742 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:02Z","lastTransitionTime":"2025-11-22T04:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.586432 4818 generic.go:334] "Generic (PLEG): container finished" podID="0f281330-68fd-4866-a787-83440a58aa4b" containerID="c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8" exitCode=0 Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.586544 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" event={"ID":"0f281330-68fd-4866-a787-83440a58aa4b","Type":"ContainerDied","Data":"c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8"} Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.586688 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.599525 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:02Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.614200 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:02Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.641131 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://120854347bb61d6824e445e5af011b9a91988e9334cd9b0855c9e8a9282028cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:02Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.658272 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:02Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.666681 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.666739 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.666756 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.666777 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.666794 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:02Z","lastTransitionTime":"2025-11-22T04:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.670461 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:02Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.685128 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:02Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.701208 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:02Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.713784 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:02Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.726442 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:02Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.739995 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:02Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.754676 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:02Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.767023 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:02Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.769015 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.769058 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.769069 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.769085 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.769097 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:02Z","lastTransitionTime":"2025-11-22T04:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.777314 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:02Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.874852 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.874888 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.874899 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.874938 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.874953 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:02Z","lastTransitionTime":"2025-11-22T04:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.978223 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.978322 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.978340 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.978361 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:02 crc kubenswrapper[4818]: I1122 04:48:02.978375 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:02Z","lastTransitionTime":"2025-11-22T04:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.080848 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.080885 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.080896 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.080912 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.080926 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:03Z","lastTransitionTime":"2025-11-22T04:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.183691 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.183737 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.183755 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.183776 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.183792 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:03Z","lastTransitionTime":"2025-11-22T04:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.286932 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.286977 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.286994 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.287016 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.287033 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:03Z","lastTransitionTime":"2025-11-22T04:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.293473 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.293566 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:03 crc kubenswrapper[4818]: E1122 04:48:03.293626 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:03 crc kubenswrapper[4818]: E1122 04:48:03.293789 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.389796 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.389851 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.389865 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.389883 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.389894 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:03Z","lastTransitionTime":"2025-11-22T04:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.492741 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.492794 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.492812 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.492834 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.492854 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:03Z","lastTransitionTime":"2025-11-22T04:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.594019 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.594039 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" event={"ID":"0f281330-68fd-4866-a787-83440a58aa4b","Type":"ContainerStarted","Data":"848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b"} Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.594991 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.595073 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.595093 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.595116 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.595135 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:03Z","lastTransitionTime":"2025-11-22T04:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.620885 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:03Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.637404 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:03Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.649958 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:03Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.666484 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:03Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.682495 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:03Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.697691 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:03Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.698701 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.698744 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.698758 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.698775 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.698787 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:03Z","lastTransitionTime":"2025-11-22T04:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.714708 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:03Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.728319 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:03Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.742306 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:03Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.758580 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:03Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.779256 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://120854347bb61d6824e445e5af011b9a91988e9334cd9b0855c9e8a9282028cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:03Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.795402 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:03Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.800690 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.800730 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.800743 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.800759 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.800772 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:03Z","lastTransitionTime":"2025-11-22T04:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.804815 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:03Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.903589 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.903630 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.903641 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.903657 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:03 crc kubenswrapper[4818]: I1122 04:48:03.903669 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:03Z","lastTransitionTime":"2025-11-22T04:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.006545 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.006595 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.006610 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.006632 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.006647 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:04Z","lastTransitionTime":"2025-11-22T04:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.109968 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.110031 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.110048 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.110070 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.110089 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:04Z","lastTransitionTime":"2025-11-22T04:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.212408 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.212452 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.212464 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.212480 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.212491 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:04Z","lastTransitionTime":"2025-11-22T04:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.290873 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:04 crc kubenswrapper[4818]: E1122 04:48:04.291150 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.315712 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.315781 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.315799 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.315821 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.315837 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:04Z","lastTransitionTime":"2025-11-22T04:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.418465 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.418560 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.418584 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.418615 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.418637 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:04Z","lastTransitionTime":"2025-11-22T04:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.446779 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk"] Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.447604 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.456856 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.458026 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.470869 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.486639 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.505197 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://120854347bb61d6824e445e5af011b9a91988e9334cd9b0855c9e8a9282028cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.521748 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.521820 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.521843 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.521878 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.521904 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:04Z","lastTransitionTime":"2025-11-22T04:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.526402 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.539418 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.555245 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.571143 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.573430 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl2rn\" (UniqueName: \"kubernetes.io/projected/ff0fda14-7029-4dfe-adf4-7b3301d926b0-kube-api-access-kl2rn\") pod \"ovnkube-control-plane-749d76644c-x52mk\" (UID: \"ff0fda14-7029-4dfe-adf4-7b3301d926b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.573490 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ff0fda14-7029-4dfe-adf4-7b3301d926b0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-x52mk\" (UID: \"ff0fda14-7029-4dfe-adf4-7b3301d926b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.573548 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ff0fda14-7029-4dfe-adf4-7b3301d926b0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-x52mk\" (UID: \"ff0fda14-7029-4dfe-adf4-7b3301d926b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.573668 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ff0fda14-7029-4dfe-adf4-7b3301d926b0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-x52mk\" (UID: \"ff0fda14-7029-4dfe-adf4-7b3301d926b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.592112 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.600381 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p8s7r_f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/ovnkube-controller/0.log" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.605028 4818 generic.go:334] "Generic (PLEG): container finished" podID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerID="120854347bb61d6824e445e5af011b9a91988e9334cd9b0855c9e8a9282028cb" exitCode=1 Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.605074 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerDied","Data":"120854347bb61d6824e445e5af011b9a91988e9334cd9b0855c9e8a9282028cb"} Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.606333 4818 scope.go:117] "RemoveContainer" containerID="120854347bb61d6824e445e5af011b9a91988e9334cd9b0855c9e8a9282028cb" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.625081 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.625132 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.625161 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.625178 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.625193 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:04Z","lastTransitionTime":"2025-11-22T04:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.645808 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.658320 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.675855 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ff0fda14-7029-4dfe-adf4-7b3301d926b0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-x52mk\" (UID: \"ff0fda14-7029-4dfe-adf4-7b3301d926b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.676050 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl2rn\" (UniqueName: \"kubernetes.io/projected/ff0fda14-7029-4dfe-adf4-7b3301d926b0-kube-api-access-kl2rn\") pod \"ovnkube-control-plane-749d76644c-x52mk\" (UID: \"ff0fda14-7029-4dfe-adf4-7b3301d926b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.677063 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ff0fda14-7029-4dfe-adf4-7b3301d926b0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-x52mk\" (UID: \"ff0fda14-7029-4dfe-adf4-7b3301d926b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.677790 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.678157 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ff0fda14-7029-4dfe-adf4-7b3301d926b0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-x52mk\" (UID: \"ff0fda14-7029-4dfe-adf4-7b3301d926b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.678497 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ff0fda14-7029-4dfe-adf4-7b3301d926b0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-x52mk\" (UID: \"ff0fda14-7029-4dfe-adf4-7b3301d926b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.678941 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ff0fda14-7029-4dfe-adf4-7b3301d926b0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-x52mk\" (UID: \"ff0fda14-7029-4dfe-adf4-7b3301d926b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.687037 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ff0fda14-7029-4dfe-adf4-7b3301d926b0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-x52mk\" (UID: \"ff0fda14-7029-4dfe-adf4-7b3301d926b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.694211 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl2rn\" (UniqueName: \"kubernetes.io/projected/ff0fda14-7029-4dfe-adf4-7b3301d926b0-kube-api-access-kl2rn\") pod \"ovnkube-control-plane-749d76644c-x52mk\" (UID: \"ff0fda14-7029-4dfe-adf4-7b3301d926b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.699542 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.718442 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.736704 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.736761 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.736776 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.736795 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.736836 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:04Z","lastTransitionTime":"2025-11-22T04:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.737538 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.755839 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.768310 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.776191 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: W1122 04:48:04.790925 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff0fda14_7029_4dfe_adf4_7b3301d926b0.slice/crio-1f1fb25dd870afbfb33065dd821dfc430a5d660c2e7ba30be738236ebffa3deb WatchSource:0}: Error finding container 1f1fb25dd870afbfb33065dd821dfc430a5d660c2e7ba30be738236ebffa3deb: Status 404 returned error can't find the container with id 1f1fb25dd870afbfb33065dd821dfc430a5d660c2e7ba30be738236ebffa3deb Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.791107 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.809156 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.828217 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.839963 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.840317 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.840333 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.840351 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.840364 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:04Z","lastTransitionTime":"2025-11-22T04:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.847619 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.869020 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://120854347bb61d6824e445e5af011b9a91988e9334cd9b0855c9e8a9282028cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://120854347bb61d6824e445e5af011b9a91988e9334cd9b0855c9e8a9282028cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"message\\\":\\\"er.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 04:48:03.766007 6071 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1122 04:48:03.766046 6071 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1122 04:48:03.766071 6071 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 04:48:03.766174 6071 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:03.766351 6071 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:03.766552 6071 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:03.767321 6071 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1122 04:48:03.767346 6071 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1122 04:48:03.767375 6071 factory.go:656] Stopping watch factory\\\\nI1122 04:48:03.767385 6071 ovnkube.go:599] Stopped ovnkube\\\\nI1122 04:48:03.767392 6071 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 04:48:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.885381 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.904109 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.917205 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.928583 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.942820 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.942861 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.942875 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.942895 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.942907 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:04Z","lastTransitionTime":"2025-11-22T04:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.946627 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.963560 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:04 crc kubenswrapper[4818]: I1122 04:48:04.979240 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:04Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.035144 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.046206 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.046243 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.046255 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.046289 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.046301 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:05Z","lastTransitionTime":"2025-11-22T04:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.148515 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.148561 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.148570 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.148583 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.148595 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:05Z","lastTransitionTime":"2025-11-22T04:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.251208 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.251300 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.251318 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.251341 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.251360 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:05Z","lastTransitionTime":"2025-11-22T04:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.290915 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.291003 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:05 crc kubenswrapper[4818]: E1122 04:48:05.291093 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:05 crc kubenswrapper[4818]: E1122 04:48:05.291186 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.354485 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.354560 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.354586 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.354615 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.354643 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:05Z","lastTransitionTime":"2025-11-22T04:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.458225 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.458357 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.458382 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.458413 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.458438 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:05Z","lastTransitionTime":"2025-11-22T04:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.562548 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.562594 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.562610 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.562631 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.562648 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:05Z","lastTransitionTime":"2025-11-22T04:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.621721 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p8s7r_f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/ovnkube-controller/0.log" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.626708 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerStarted","Data":"e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832"} Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.633694 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" event={"ID":"ff0fda14-7029-4dfe-adf4-7b3301d926b0","Type":"ContainerStarted","Data":"1f1fb25dd870afbfb33065dd821dfc430a5d660c2e7ba30be738236ebffa3deb"} Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.665655 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.665903 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.666009 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.666144 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.666310 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:05Z","lastTransitionTime":"2025-11-22T04:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.770072 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.770128 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.770146 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.770168 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.770184 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:05Z","lastTransitionTime":"2025-11-22T04:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.872984 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.873037 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.873049 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.873069 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.873081 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:05Z","lastTransitionTime":"2025-11-22T04:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.976637 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.976692 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.976745 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.976769 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:05 crc kubenswrapper[4818]: I1122 04:48:05.976787 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:05Z","lastTransitionTime":"2025-11-22T04:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.080142 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.080218 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.080238 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.080308 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.080333 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:06Z","lastTransitionTime":"2025-11-22T04:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.183611 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.183936 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.184106 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.184307 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.184499 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:06Z","lastTransitionTime":"2025-11-22T04:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.287839 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.287939 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.287958 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.287983 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.288002 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:06Z","lastTransitionTime":"2025-11-22T04:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.290293 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:06 crc kubenswrapper[4818]: E1122 04:48:06.290480 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.357070 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-djd77"] Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.357915 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:06 crc kubenswrapper[4818]: E1122 04:48:06.358045 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.374941 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.390291 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.390348 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.390361 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.390392 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.390408 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:06Z","lastTransitionTime":"2025-11-22T04:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.391414 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.415428 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.437455 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.456326 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.471533 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.487441 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.496768 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.496831 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.496845 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.496871 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.496889 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:06Z","lastTransitionTime":"2025-11-22T04:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.502630 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd652\" (UniqueName: \"kubernetes.io/projected/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-kube-api-access-bd652\") pod \"network-metrics-daemon-djd77\" (UID: \"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\") " pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.502710 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs\") pod \"network-metrics-daemon-djd77\" (UID: \"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\") " pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.504650 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.523186 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.541224 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.558713 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.574879 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.591600 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.599442 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.599496 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.599507 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.599527 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.599542 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:06Z","lastTransitionTime":"2025-11-22T04:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.603818 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs\") pod \"network-metrics-daemon-djd77\" (UID: \"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\") " pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.603896 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd652\" (UniqueName: \"kubernetes.io/projected/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-kube-api-access-bd652\") pod \"network-metrics-daemon-djd77\" (UID: \"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\") " pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:06 crc kubenswrapper[4818]: E1122 04:48:06.603955 4818 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 04:48:06 crc kubenswrapper[4818]: E1122 04:48:06.604032 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs podName:429a37e7-cb52-4e76-89ee-16f2b9e6fbfe nodeName:}" failed. No retries permitted until 2025-11-22 04:48:07.10401142 +0000 UTC m=+39.678427967 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs") pod "network-metrics-daemon-djd77" (UID: "429a37e7-cb52-4e76-89ee-16f2b9e6fbfe") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.616929 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://120854347bb61d6824e445e5af011b9a91988e9334cd9b0855c9e8a9282028cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://120854347bb61d6824e445e5af011b9a91988e9334cd9b0855c9e8a9282028cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"message\\\":\\\"er.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 04:48:03.766007 6071 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1122 04:48:03.766046 6071 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1122 04:48:03.766071 6071 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 04:48:03.766174 6071 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:03.766351 6071 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:03.766552 6071 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:03.767321 6071 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1122 04:48:03.767346 6071 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1122 04:48:03.767375 6071 factory.go:656] Stopping watch factory\\\\nI1122 04:48:03.767385 6071 ovnkube.go:599] Stopped ovnkube\\\\nI1122 04:48:03.767392 6071 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 04:48:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.621218 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd652\" (UniqueName: \"kubernetes.io/projected/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-kube-api-access-bd652\") pod \"network-metrics-daemon-djd77\" (UID: \"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\") " pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.633695 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.637946 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" event={"ID":"ff0fda14-7029-4dfe-adf4-7b3301d926b0","Type":"ContainerStarted","Data":"2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d"} Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.638300 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.660291 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.674331 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.687753 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.698603 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.701587 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.701611 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.701618 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.701631 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.701639 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:06Z","lastTransitionTime":"2025-11-22T04:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.708549 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.721732 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.741362 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.761005 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://120854347bb61d6824e445e5af011b9a91988e9334cd9b0855c9e8a9282028cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"message\\\":\\\"er.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 04:48:03.766007 6071 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1122 04:48:03.766046 6071 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1122 04:48:03.766071 6071 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 04:48:03.766174 6071 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:03.766351 6071 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:03.766552 6071 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:03.767321 6071 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1122 04:48:03.767346 6071 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1122 04:48:03.767375 6071 factory.go:656] Stopping watch factory\\\\nI1122 04:48:03.767385 6071 ovnkube.go:599] Stopped ovnkube\\\\nI1122 04:48:03.767392 6071 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 04:48:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.789334 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.804152 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.804180 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.804188 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.804200 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.804210 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:06Z","lastTransitionTime":"2025-11-22T04:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.815002 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.844981 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.854802 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.867721 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.882463 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.895610 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:06Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.906641 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.906683 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.906694 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.906710 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:06 crc kubenswrapper[4818]: I1122 04:48:06.906723 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:06Z","lastTransitionTime":"2025-11-22T04:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.008694 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.008750 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.008763 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.008780 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.008795 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:07Z","lastTransitionTime":"2025-11-22T04:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.107442 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs\") pod \"network-metrics-daemon-djd77\" (UID: \"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\") " pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:07 crc kubenswrapper[4818]: E1122 04:48:07.107659 4818 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 04:48:07 crc kubenswrapper[4818]: E1122 04:48:07.107828 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs podName:429a37e7-cb52-4e76-89ee-16f2b9e6fbfe nodeName:}" failed. No retries permitted until 2025-11-22 04:48:08.10779784 +0000 UTC m=+40.682214397 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs") pod "network-metrics-daemon-djd77" (UID: "429a37e7-cb52-4e76-89ee-16f2b9e6fbfe") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.111450 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.111494 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.111506 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.111524 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.111536 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:07Z","lastTransitionTime":"2025-11-22T04:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.214718 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.214776 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.214799 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.214828 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.214849 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:07Z","lastTransitionTime":"2025-11-22T04:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.290539 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.290631 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:07 crc kubenswrapper[4818]: E1122 04:48:07.290723 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:07 crc kubenswrapper[4818]: E1122 04:48:07.290861 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.317642 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.317706 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.317729 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.317759 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.317781 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:07Z","lastTransitionTime":"2025-11-22T04:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.421110 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.421150 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.421158 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.421173 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.421182 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:07Z","lastTransitionTime":"2025-11-22T04:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.523400 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.523453 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.523462 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.523474 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.523483 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:07Z","lastTransitionTime":"2025-11-22T04:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.626471 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.626519 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.626531 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.626547 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.626559 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:07Z","lastTransitionTime":"2025-11-22T04:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.644073 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p8s7r_f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/ovnkube-controller/1.log" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.644874 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p8s7r_f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/ovnkube-controller/0.log" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.648293 4818 generic.go:334] "Generic (PLEG): container finished" podID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerID="e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832" exitCode=1 Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.648349 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerDied","Data":"e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832"} Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.648413 4818 scope.go:117] "RemoveContainer" containerID="120854347bb61d6824e445e5af011b9a91988e9334cd9b0855c9e8a9282028cb" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.649657 4818 scope.go:117] "RemoveContainer" containerID="e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832" Nov 22 04:48:07 crc kubenswrapper[4818]: E1122 04:48:07.649880 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.653402 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" event={"ID":"ff0fda14-7029-4dfe-adf4-7b3301d926b0","Type":"ContainerStarted","Data":"9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049"} Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.673995 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:07Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.697285 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:07Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.713966 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:07Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.729973 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.730037 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.730054 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.730078 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.730095 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:07Z","lastTransitionTime":"2025-11-22T04:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.731485 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:07Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.751614 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:07Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.774361 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:07Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.794505 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:07Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.815882 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:07Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.832681 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.832756 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.832769 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.832785 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.832816 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:07Z","lastTransitionTime":"2025-11-22T04:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.838400 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:07Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.860654 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://120854347bb61d6824e445e5af011b9a91988e9334cd9b0855c9e8a9282028cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"message\\\":\\\"er.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 04:48:03.766007 6071 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1122 04:48:03.766046 6071 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1122 04:48:03.766071 6071 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 04:48:03.766174 6071 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:03.766351 6071 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:03.766552 6071 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:03.767321 6071 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1122 04:48:03.767346 6071 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1122 04:48:03.767375 6071 factory.go:656] Stopping watch factory\\\\nI1122 04:48:03.767385 6071 ovnkube.go:599] Stopped ovnkube\\\\nI1122 04:48:03.767392 6071 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 04:48:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"message\\\":\\\"ubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:07.148568 6258 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1122 04:48:07.148772 6258 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:07.149004 6258 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:07.149078 6258 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:07.194406 6258 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1122 04:48:07.194444 6258 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1122 04:48:07.194523 6258 ovnkube.go:599] Stopped ovnkube\\\\nI1122 04:48:07.194561 6258 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 04:48:07.194666 6258 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:07Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.880203 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:07Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.895918 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:07Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.913541 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:07Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.931740 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:07Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.935730 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.935786 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.935807 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.935833 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.935850 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:07Z","lastTransitionTime":"2025-11-22T04:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.949150 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:07Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.964009 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:07Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.979612 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:07Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:07 crc kubenswrapper[4818]: I1122 04:48:07.998348 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:07Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.017052 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.032315 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.038001 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.038069 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.038087 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.038112 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.038128 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:08Z","lastTransitionTime":"2025-11-22T04:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.046424 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.061974 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.075859 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.087722 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.104213 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.118312 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs\") pod \"network-metrics-daemon-djd77\" (UID: \"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\") " pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:08 crc kubenswrapper[4818]: E1122 04:48:08.118466 4818 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 04:48:08 crc kubenswrapper[4818]: E1122 04:48:08.118526 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs podName:429a37e7-cb52-4e76-89ee-16f2b9e6fbfe nodeName:}" failed. No retries permitted until 2025-11-22 04:48:10.118507942 +0000 UTC m=+42.692924479 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs") pod "network-metrics-daemon-djd77" (UID: "429a37e7-cb52-4e76-89ee-16f2b9e6fbfe") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.125154 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.141406 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.141446 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.141455 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.141468 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.141479 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:08Z","lastTransitionTime":"2025-11-22T04:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.159957 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://120854347bb61d6824e445e5af011b9a91988e9334cd9b0855c9e8a9282028cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"message\\\":\\\"er.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 04:48:03.766007 6071 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1122 04:48:03.766046 6071 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1122 04:48:03.766071 6071 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 04:48:03.766174 6071 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:03.766351 6071 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:03.766552 6071 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:03.767321 6071 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1122 04:48:03.767346 6071 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1122 04:48:03.767375 6071 factory.go:656] Stopping watch factory\\\\nI1122 04:48:03.767385 6071 ovnkube.go:599] Stopped ovnkube\\\\nI1122 04:48:03.767392 6071 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 04:48:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"message\\\":\\\"ubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:07.148568 6258 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1122 04:48:07.148772 6258 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:07.149004 6258 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:07.149078 6258 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:07.194406 6258 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1122 04:48:07.194444 6258 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1122 04:48:07.194523 6258 ovnkube.go:599] Stopped ovnkube\\\\nI1122 04:48:07.194561 6258 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 04:48:07.194666 6258 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.176075 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.186613 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.203218 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.244037 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.244085 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.244097 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.244150 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.244171 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:08Z","lastTransitionTime":"2025-11-22T04:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.290602 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:08 crc kubenswrapper[4818]: E1122 04:48:08.290761 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.290889 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:08 crc kubenswrapper[4818]: E1122 04:48:08.291326 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.306008 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.324340 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.346692 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.346737 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.346747 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.346764 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.346774 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:08Z","lastTransitionTime":"2025-11-22T04:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.352182 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://120854347bb61d6824e445e5af011b9a91988e9334cd9b0855c9e8a9282028cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"message\\\":\\\"er.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 04:48:03.766007 6071 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1122 04:48:03.766046 6071 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1122 04:48:03.766071 6071 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 04:48:03.766174 6071 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:03.766351 6071 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:03.766552 6071 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:03.767321 6071 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1122 04:48:03.767346 6071 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1122 04:48:03.767375 6071 factory.go:656] Stopping watch factory\\\\nI1122 04:48:03.767385 6071 ovnkube.go:599] Stopped ovnkube\\\\nI1122 04:48:03.767392 6071 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 04:48:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"message\\\":\\\"ubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:07.148568 6258 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1122 04:48:07.148772 6258 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:07.149004 6258 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:07.149078 6258 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:07.194406 6258 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1122 04:48:07.194444 6258 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1122 04:48:07.194523 6258 ovnkube.go:599] Stopped ovnkube\\\\nI1122 04:48:07.194561 6258 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 04:48:07.194666 6258 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.369053 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.380748 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.390515 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.416178 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.434282 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.448547 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.448612 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.448630 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.448654 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.448672 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:08Z","lastTransitionTime":"2025-11-22T04:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.450077 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.467074 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.487181 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.506897 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.528353 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.544613 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.552847 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.552971 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.553052 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.553135 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.553218 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:08Z","lastTransitionTime":"2025-11-22T04:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.556632 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.656330 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.656397 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.656411 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.656434 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.656451 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:08Z","lastTransitionTime":"2025-11-22T04:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.660527 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p8s7r_f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/ovnkube-controller/1.log" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.665304 4818 scope.go:117] "RemoveContainer" containerID="e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832" Nov 22 04:48:08 crc kubenswrapper[4818]: E1122 04:48:08.665578 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.679111 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.692287 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.706433 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.720471 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.737385 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.756572 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.759412 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.759439 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.759452 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.759473 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.759486 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:08Z","lastTransitionTime":"2025-11-22T04:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.773140 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.796276 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.809901 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.826165 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.853224 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"message\\\":\\\"ubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:07.148568 6258 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1122 04:48:07.148772 6258 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:07.149004 6258 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:07.149078 6258 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:07.194406 6258 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1122 04:48:07.194444 6258 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1122 04:48:07.194523 6258 ovnkube.go:599] Stopped ovnkube\\\\nI1122 04:48:07.194561 6258 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 04:48:07.194666 6258 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.862726 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.862793 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.862812 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.862837 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.862857 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:08Z","lastTransitionTime":"2025-11-22T04:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.870061 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.889182 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.913768 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.936637 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.966090 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.966152 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.966173 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.966201 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:08 crc kubenswrapper[4818]: I1122 04:48:08.966224 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:08Z","lastTransitionTime":"2025-11-22T04:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.069586 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.069654 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.069671 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.069695 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.069712 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:09Z","lastTransitionTime":"2025-11-22T04:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.131455 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.131728 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:09 crc kubenswrapper[4818]: E1122 04:48:09.131771 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:48:25.131732118 +0000 UTC m=+57.706148685 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.131882 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:09 crc kubenswrapper[4818]: E1122 04:48:09.131900 4818 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 04:48:09 crc kubenswrapper[4818]: E1122 04:48:09.131985 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 04:48:25.131962864 +0000 UTC m=+57.706379431 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 04:48:09 crc kubenswrapper[4818]: E1122 04:48:09.132163 4818 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 04:48:09 crc kubenswrapper[4818]: E1122 04:48:09.132298 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 04:48:25.132233591 +0000 UTC m=+57.706650148 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.172680 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.173006 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.173133 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.173330 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.173536 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:09Z","lastTransitionTime":"2025-11-22T04:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.233015 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.233099 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:09 crc kubenswrapper[4818]: E1122 04:48:09.233332 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 04:48:09 crc kubenswrapper[4818]: E1122 04:48:09.233361 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 04:48:09 crc kubenswrapper[4818]: E1122 04:48:09.233379 4818 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:48:09 crc kubenswrapper[4818]: E1122 04:48:09.233444 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 04:48:25.23342326 +0000 UTC m=+57.807839827 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:48:09 crc kubenswrapper[4818]: E1122 04:48:09.233913 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 04:48:09 crc kubenswrapper[4818]: E1122 04:48:09.233936 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 04:48:09 crc kubenswrapper[4818]: E1122 04:48:09.233953 4818 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:48:09 crc kubenswrapper[4818]: E1122 04:48:09.234001 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 04:48:25.233986045 +0000 UTC m=+57.808402602 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.276873 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.277076 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.277100 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.277125 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.277144 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:09Z","lastTransitionTime":"2025-11-22T04:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.290467 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.290483 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:09 crc kubenswrapper[4818]: E1122 04:48:09.290610 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:09 crc kubenswrapper[4818]: E1122 04:48:09.290721 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.380130 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.380208 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.380233 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.380294 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.380320 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:09Z","lastTransitionTime":"2025-11-22T04:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.484855 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.484922 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.484946 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.484974 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.484995 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:09Z","lastTransitionTime":"2025-11-22T04:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.588423 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.588481 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.588501 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.588529 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.588552 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:09Z","lastTransitionTime":"2025-11-22T04:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.691408 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.691477 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.691501 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.691529 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.691552 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:09Z","lastTransitionTime":"2025-11-22T04:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.794935 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.795008 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.795031 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.795067 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.795089 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:09Z","lastTransitionTime":"2025-11-22T04:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.898770 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.898822 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.898840 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.898861 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:09 crc kubenswrapper[4818]: I1122 04:48:09.898880 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:09Z","lastTransitionTime":"2025-11-22T04:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.001944 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.002017 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.002037 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.002061 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.002078 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:10Z","lastTransitionTime":"2025-11-22T04:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.104777 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.104843 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.105034 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.105072 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.105094 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:10Z","lastTransitionTime":"2025-11-22T04:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.143178 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs\") pod \"network-metrics-daemon-djd77\" (UID: \"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\") " pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:10 crc kubenswrapper[4818]: E1122 04:48:10.143402 4818 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 04:48:10 crc kubenswrapper[4818]: E1122 04:48:10.143504 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs podName:429a37e7-cb52-4e76-89ee-16f2b9e6fbfe nodeName:}" failed. No retries permitted until 2025-11-22 04:48:14.143478027 +0000 UTC m=+46.717894584 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs") pod "network-metrics-daemon-djd77" (UID: "429a37e7-cb52-4e76-89ee-16f2b9e6fbfe") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.208347 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.208426 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.208444 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.208467 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.208485 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:10Z","lastTransitionTime":"2025-11-22T04:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.290484 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.290524 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:10 crc kubenswrapper[4818]: E1122 04:48:10.290688 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:10 crc kubenswrapper[4818]: E1122 04:48:10.290893 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.311762 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.311840 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.311866 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.311901 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.311925 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:10Z","lastTransitionTime":"2025-11-22T04:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.415549 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.415612 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.415632 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.415660 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.415678 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:10Z","lastTransitionTime":"2025-11-22T04:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.519716 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.519777 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.519791 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.519810 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.519822 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:10Z","lastTransitionTime":"2025-11-22T04:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.623900 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.623976 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.624003 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.624033 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.624056 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:10Z","lastTransitionTime":"2025-11-22T04:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.732474 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.732932 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.732953 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.732978 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.732995 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:10Z","lastTransitionTime":"2025-11-22T04:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.835462 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.835525 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.835542 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.835566 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.835583 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:10Z","lastTransitionTime":"2025-11-22T04:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.938606 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.938679 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.938702 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.938735 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:10 crc kubenswrapper[4818]: I1122 04:48:10.938758 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:10Z","lastTransitionTime":"2025-11-22T04:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.041663 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.041758 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.041777 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.041800 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.041817 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:11Z","lastTransitionTime":"2025-11-22T04:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.145950 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.146006 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.146022 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.146045 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.146062 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:11Z","lastTransitionTime":"2025-11-22T04:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.248587 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.248628 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.248637 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.248650 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.248659 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:11Z","lastTransitionTime":"2025-11-22T04:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.290521 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:11 crc kubenswrapper[4818]: E1122 04:48:11.290710 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.291301 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:11 crc kubenswrapper[4818]: E1122 04:48:11.291454 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.351438 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.351495 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.351511 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.351534 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.351551 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:11Z","lastTransitionTime":"2025-11-22T04:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.454494 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.454544 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.454563 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.454586 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.454602 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:11Z","lastTransitionTime":"2025-11-22T04:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.481964 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.482043 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.482068 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.482104 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.482128 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:11Z","lastTransitionTime":"2025-11-22T04:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:11 crc kubenswrapper[4818]: E1122 04:48:11.498683 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:11Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.504659 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.504730 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.504746 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.504772 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.504790 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:11Z","lastTransitionTime":"2025-11-22T04:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:11 crc kubenswrapper[4818]: E1122 04:48:11.523548 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:11Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.527638 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.527704 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.527723 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.527749 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.527767 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:11Z","lastTransitionTime":"2025-11-22T04:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:11 crc kubenswrapper[4818]: E1122 04:48:11.548475 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:11Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.552821 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.552859 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.552867 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.552883 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.552895 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:11Z","lastTransitionTime":"2025-11-22T04:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:11 crc kubenswrapper[4818]: E1122 04:48:11.572001 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:11Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.576677 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.576720 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.576738 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.576765 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.576784 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:11Z","lastTransitionTime":"2025-11-22T04:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:11 crc kubenswrapper[4818]: E1122 04:48:11.596559 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:11Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:11 crc kubenswrapper[4818]: E1122 04:48:11.596809 4818 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.599563 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.599634 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.599656 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.599690 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.599730 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:11Z","lastTransitionTime":"2025-11-22T04:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.702928 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.703004 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.703022 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.703049 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.703068 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:11Z","lastTransitionTime":"2025-11-22T04:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.807056 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.807133 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.807162 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.807194 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.807213 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:11Z","lastTransitionTime":"2025-11-22T04:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.911029 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.911103 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.911127 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.911153 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:11 crc kubenswrapper[4818]: I1122 04:48:11.911175 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:11Z","lastTransitionTime":"2025-11-22T04:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.014395 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.014480 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.014512 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.014542 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.014562 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:12Z","lastTransitionTime":"2025-11-22T04:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.118159 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.118227 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.118244 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.118314 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.118332 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:12Z","lastTransitionTime":"2025-11-22T04:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.222011 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.222306 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.222326 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.222352 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.222370 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:12Z","lastTransitionTime":"2025-11-22T04:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.290613 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:12 crc kubenswrapper[4818]: E1122 04:48:12.290799 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.290889 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:12 crc kubenswrapper[4818]: E1122 04:48:12.291100 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.325368 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.325442 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.325456 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.325479 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.325495 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:12Z","lastTransitionTime":"2025-11-22T04:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.428668 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.428741 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.428766 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.428797 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.428820 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:12Z","lastTransitionTime":"2025-11-22T04:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.532515 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.532577 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.532594 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.532618 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.532636 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:12Z","lastTransitionTime":"2025-11-22T04:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.636606 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.636699 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.636730 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.636762 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.636782 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:12Z","lastTransitionTime":"2025-11-22T04:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.739752 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.739804 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.739817 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.739842 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.739860 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:12Z","lastTransitionTime":"2025-11-22T04:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.842667 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.842715 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.842727 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.842743 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.842755 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:12Z","lastTransitionTime":"2025-11-22T04:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.946355 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.946422 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.946442 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.946467 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:12 crc kubenswrapper[4818]: I1122 04:48:12.946485 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:12Z","lastTransitionTime":"2025-11-22T04:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.048955 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.049006 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.049017 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.049034 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.049045 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:13Z","lastTransitionTime":"2025-11-22T04:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.151822 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.151880 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.151897 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.151921 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.151937 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:13Z","lastTransitionTime":"2025-11-22T04:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.254537 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.254591 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.254609 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.254631 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.254647 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:13Z","lastTransitionTime":"2025-11-22T04:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.290370 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:13 crc kubenswrapper[4818]: E1122 04:48:13.290516 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.290590 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:13 crc kubenswrapper[4818]: E1122 04:48:13.290760 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.291933 4818 scope.go:117] "RemoveContainer" containerID="e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.356974 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.357003 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.357017 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.357033 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.357044 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:13Z","lastTransitionTime":"2025-11-22T04:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.459229 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.459334 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.459353 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.459379 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.459397 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:13Z","lastTransitionTime":"2025-11-22T04:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.562876 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.562933 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.562951 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.562975 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.562993 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:13Z","lastTransitionTime":"2025-11-22T04:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.666030 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.666065 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.666074 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.666087 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.666098 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:13Z","lastTransitionTime":"2025-11-22T04:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.691910 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.693961 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c569fd0932bade53c98da764d0d0e9dd07755d73dbba866c0be0209661aaa05a"} Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.694504 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.709783 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:13Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.725706 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:13Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.744638 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"message\\\":\\\"ubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:07.148568 6258 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1122 04:48:07.148772 6258 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:07.149004 6258 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:07.149078 6258 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:07.194406 6258 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1122 04:48:07.194444 6258 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1122 04:48:07.194523 6258 ovnkube.go:599] Stopped ovnkube\\\\nI1122 04:48:07.194561 6258 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 04:48:07.194666 6258 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:13Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.764732 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:13Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.768801 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.768849 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.768861 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.768885 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.768900 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:13Z","lastTransitionTime":"2025-11-22T04:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.778504 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:13Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.792957 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:13Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.810462 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c569fd0932bade53c98da764d0d0e9dd07755d73dbba866c0be0209661aaa05a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:13Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.824443 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:13Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.835010 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:13Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.848451 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:13Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.867336 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:13Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.872028 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.872062 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.872070 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.872084 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.872093 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:13Z","lastTransitionTime":"2025-11-22T04:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.880498 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:13Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.890666 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:13Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.900319 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:13Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.910022 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:13Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.974923 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.974959 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.974968 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.974981 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:13 crc kubenswrapper[4818]: I1122 04:48:13.974992 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:13Z","lastTransitionTime":"2025-11-22T04:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.077644 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.077713 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.077735 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.077765 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.077789 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:14Z","lastTransitionTime":"2025-11-22T04:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.180628 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.180688 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.180706 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.180730 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.180750 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:14Z","lastTransitionTime":"2025-11-22T04:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.192690 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs\") pod \"network-metrics-daemon-djd77\" (UID: \"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\") " pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:14 crc kubenswrapper[4818]: E1122 04:48:14.192939 4818 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 04:48:14 crc kubenswrapper[4818]: E1122 04:48:14.193106 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs podName:429a37e7-cb52-4e76-89ee-16f2b9e6fbfe nodeName:}" failed. No retries permitted until 2025-11-22 04:48:22.193065418 +0000 UTC m=+54.767481985 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs") pod "network-metrics-daemon-djd77" (UID: "429a37e7-cb52-4e76-89ee-16f2b9e6fbfe") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.283850 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.283910 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.283927 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.283951 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.283968 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:14Z","lastTransitionTime":"2025-11-22T04:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.290181 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.290186 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:14 crc kubenswrapper[4818]: E1122 04:48:14.290428 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:14 crc kubenswrapper[4818]: E1122 04:48:14.290539 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.387407 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.387470 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.387487 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.387511 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.387533 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:14Z","lastTransitionTime":"2025-11-22T04:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.490091 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.490128 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.490138 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.490151 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.490162 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:14Z","lastTransitionTime":"2025-11-22T04:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.592573 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.592641 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.592661 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.592688 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.592709 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:14Z","lastTransitionTime":"2025-11-22T04:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.696396 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.696505 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.696549 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.696605 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.696624 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:14Z","lastTransitionTime":"2025-11-22T04:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.800449 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.800511 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.800551 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.800584 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.800607 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:14Z","lastTransitionTime":"2025-11-22T04:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.903706 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.903783 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.903802 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.903832 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:14 crc kubenswrapper[4818]: I1122 04:48:14.903852 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:14Z","lastTransitionTime":"2025-11-22T04:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.006441 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.006498 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.006509 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.006527 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.006538 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:15Z","lastTransitionTime":"2025-11-22T04:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.108704 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.108766 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.108791 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.108822 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.108845 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:15Z","lastTransitionTime":"2025-11-22T04:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.211040 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.211092 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.211124 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.211151 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.211170 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:15Z","lastTransitionTime":"2025-11-22T04:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.291056 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.291081 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:15 crc kubenswrapper[4818]: E1122 04:48:15.291246 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:15 crc kubenswrapper[4818]: E1122 04:48:15.291437 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.313965 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.314009 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.314018 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.314032 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.314045 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:15Z","lastTransitionTime":"2025-11-22T04:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.416415 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.416487 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.416505 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.416532 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.416549 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:15Z","lastTransitionTime":"2025-11-22T04:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.518938 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.519013 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.519038 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.519071 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.519097 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:15Z","lastTransitionTime":"2025-11-22T04:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.622625 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.622698 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.622726 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.622758 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.622783 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:15Z","lastTransitionTime":"2025-11-22T04:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.730676 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.730740 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.730783 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.730816 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.730838 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:15Z","lastTransitionTime":"2025-11-22T04:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.833284 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.833317 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.833328 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.833346 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.833357 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:15Z","lastTransitionTime":"2025-11-22T04:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.936528 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.936580 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.936597 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.936622 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:15 crc kubenswrapper[4818]: I1122 04:48:15.936640 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:15Z","lastTransitionTime":"2025-11-22T04:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.039172 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.039310 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.039338 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.039368 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.039385 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:16Z","lastTransitionTime":"2025-11-22T04:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.142461 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.142531 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.142553 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.142587 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.142609 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:16Z","lastTransitionTime":"2025-11-22T04:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.245701 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.246152 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.246177 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.246202 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.246219 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:16Z","lastTransitionTime":"2025-11-22T04:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.291033 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.291035 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:16 crc kubenswrapper[4818]: E1122 04:48:16.291250 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:16 crc kubenswrapper[4818]: E1122 04:48:16.291421 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.348493 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.348791 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.348895 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.348981 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.349058 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:16Z","lastTransitionTime":"2025-11-22T04:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.451948 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.452017 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.452042 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.452070 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.452090 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:16Z","lastTransitionTime":"2025-11-22T04:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.555143 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.555844 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.556011 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.556156 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.556350 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:16Z","lastTransitionTime":"2025-11-22T04:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.659616 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.659685 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.659712 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.659742 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.659764 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:16Z","lastTransitionTime":"2025-11-22T04:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.762531 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.762828 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.762963 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.763109 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.763216 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:16Z","lastTransitionTime":"2025-11-22T04:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.867114 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.867190 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.867218 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.867289 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.867317 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:16Z","lastTransitionTime":"2025-11-22T04:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.970379 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.970434 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.970453 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.970475 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:16 crc kubenswrapper[4818]: I1122 04:48:16.970492 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:16Z","lastTransitionTime":"2025-11-22T04:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.073043 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.073105 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.073128 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.073156 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.073177 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:17Z","lastTransitionTime":"2025-11-22T04:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.176317 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.176402 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.176456 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.176492 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.176552 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:17Z","lastTransitionTime":"2025-11-22T04:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.279675 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.279711 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.279719 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.279733 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.279743 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:17Z","lastTransitionTime":"2025-11-22T04:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.290149 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.290168 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:17 crc kubenswrapper[4818]: E1122 04:48:17.290354 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:17 crc kubenswrapper[4818]: E1122 04:48:17.290451 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.382534 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.382614 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.382637 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.382670 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.382694 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:17Z","lastTransitionTime":"2025-11-22T04:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.486118 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.486165 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.486183 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.486205 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.486222 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:17Z","lastTransitionTime":"2025-11-22T04:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.589220 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.589328 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.589387 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.589424 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.589453 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:17Z","lastTransitionTime":"2025-11-22T04:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.692799 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.692876 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.692902 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.692929 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.692949 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:17Z","lastTransitionTime":"2025-11-22T04:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.795750 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.795842 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.795868 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.795904 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.795928 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:17Z","lastTransitionTime":"2025-11-22T04:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.899675 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.899741 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.899759 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.899786 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:17 crc kubenswrapper[4818]: I1122 04:48:17.899804 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:17Z","lastTransitionTime":"2025-11-22T04:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.002752 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.002819 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.002839 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.002865 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.002885 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:18Z","lastTransitionTime":"2025-11-22T04:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.106148 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.106209 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.106226 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.106284 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.106304 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:18Z","lastTransitionTime":"2025-11-22T04:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.209591 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.209671 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.209692 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.209727 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.209753 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:18Z","lastTransitionTime":"2025-11-22T04:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.292022 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:18 crc kubenswrapper[4818]: E1122 04:48:18.292312 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.292698 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:18 crc kubenswrapper[4818]: E1122 04:48:18.293137 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.312346 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.312597 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.312723 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.312853 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.312977 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:18Z","lastTransitionTime":"2025-11-22T04:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.323750 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.339929 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.361742 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"message\\\":\\\"ubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:07.148568 6258 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1122 04:48:07.148772 6258 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:07.149004 6258 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:07.149078 6258 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:07.194406 6258 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1122 04:48:07.194444 6258 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1122 04:48:07.194523 6258 ovnkube.go:599] Stopped ovnkube\\\\nI1122 04:48:07.194561 6258 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 04:48:07.194666 6258 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.380048 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.394957 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.409324 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.418670 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.418726 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.418745 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.418769 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.418787 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:18Z","lastTransitionTime":"2025-11-22T04:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.430403 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c569fd0932bade53c98da764d0d0e9dd07755d73dbba866c0be0209661aaa05a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.448640 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.460445 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.476489 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.494761 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.510676 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.522543 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.522578 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.522591 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.522610 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.522623 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:18Z","lastTransitionTime":"2025-11-22T04:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.526132 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.544634 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.560141 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.626901 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.626990 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.627015 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.627049 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.627073 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:18Z","lastTransitionTime":"2025-11-22T04:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.729455 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.729502 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.729513 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.729528 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.729539 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:18Z","lastTransitionTime":"2025-11-22T04:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.832467 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.832518 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.832529 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.832546 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.832566 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:18Z","lastTransitionTime":"2025-11-22T04:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.939527 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.939593 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.939614 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.939638 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:18 crc kubenswrapper[4818]: I1122 04:48:18.939656 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:18Z","lastTransitionTime":"2025-11-22T04:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.042944 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.044043 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.044073 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.044120 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.044136 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:19Z","lastTransitionTime":"2025-11-22T04:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.148840 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.148921 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.148943 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.148977 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.149002 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:19Z","lastTransitionTime":"2025-11-22T04:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.252354 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.252475 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.252497 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.252523 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.252542 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:19Z","lastTransitionTime":"2025-11-22T04:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.290382 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.290378 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:19 crc kubenswrapper[4818]: E1122 04:48:19.290569 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:19 crc kubenswrapper[4818]: E1122 04:48:19.290815 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.356973 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.357047 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.357064 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.357089 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.357104 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:19Z","lastTransitionTime":"2025-11-22T04:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.461180 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.461244 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.461287 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.461497 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.461515 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:19Z","lastTransitionTime":"2025-11-22T04:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.565358 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.565444 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.565484 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.565514 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.565537 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:19Z","lastTransitionTime":"2025-11-22T04:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.668861 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.668929 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.668950 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.668974 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.668995 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:19Z","lastTransitionTime":"2025-11-22T04:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.772018 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.772094 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.772126 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.772156 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.772177 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:19Z","lastTransitionTime":"2025-11-22T04:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.875999 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.876063 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.876081 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.876105 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.876123 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:19Z","lastTransitionTime":"2025-11-22T04:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.979574 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.979625 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.979636 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.979654 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:19 crc kubenswrapper[4818]: I1122 04:48:19.979664 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:19Z","lastTransitionTime":"2025-11-22T04:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.083120 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.083202 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.083215 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.083237 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.083278 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:20Z","lastTransitionTime":"2025-11-22T04:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.186807 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.187457 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.187498 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.187514 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.187528 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:20Z","lastTransitionTime":"2025-11-22T04:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.290148 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:20 crc kubenswrapper[4818]: E1122 04:48:20.290355 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.290934 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:20 crc kubenswrapper[4818]: E1122 04:48:20.291362 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.291918 4818 scope.go:117] "RemoveContainer" containerID="e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.292192 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.292393 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.292575 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.292739 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.292896 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:20Z","lastTransitionTime":"2025-11-22T04:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.396599 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.396681 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.396703 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.396731 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.396892 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:20Z","lastTransitionTime":"2025-11-22T04:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.500194 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.500403 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.500612 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.500788 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.500857 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:20Z","lastTransitionTime":"2025-11-22T04:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.603895 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.603941 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.603950 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.603963 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.603972 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:20Z","lastTransitionTime":"2025-11-22T04:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.705682 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.705735 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.705752 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.705774 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.705791 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:20Z","lastTransitionTime":"2025-11-22T04:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.722162 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p8s7r_f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/ovnkube-controller/1.log" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.724562 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerStarted","Data":"7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b"} Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.724929 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.740762 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:20Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.757885 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:20Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.776864 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:20Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.798626 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:20Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.808863 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.808937 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.808963 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.808993 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.809016 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:20Z","lastTransitionTime":"2025-11-22T04:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.824395 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:20Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.843892 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:20Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.862385 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:20Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.886854 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:20Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.909675 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:20Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.911934 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.911982 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.912026 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.912051 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.912067 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:20Z","lastTransitionTime":"2025-11-22T04:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.931735 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"message\\\":\\\"ubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:07.148568 6258 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1122 04:48:07.148772 6258 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:07.149004 6258 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:07.149078 6258 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:07.194406 6258 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1122 04:48:07.194444 6258 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1122 04:48:07.194523 6258 ovnkube.go:599] Stopped ovnkube\\\\nI1122 04:48:07.194561 6258 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 04:48:07.194666 6258 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:20Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.945292 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:20Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.955972 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:20Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.967159 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:20Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:20 crc kubenswrapper[4818]: I1122 04:48:20.985242 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c569fd0932bade53c98da764d0d0e9dd07755d73dbba866c0be0209661aaa05a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:20Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.004112 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.014429 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.014463 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.014470 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.014482 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.014493 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:21Z","lastTransitionTime":"2025-11-22T04:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.116572 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.116605 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.116615 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.116629 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.116639 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:21Z","lastTransitionTime":"2025-11-22T04:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.219783 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.219846 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.219867 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.219899 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.219925 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:21Z","lastTransitionTime":"2025-11-22T04:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.290607 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.290727 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:21 crc kubenswrapper[4818]: E1122 04:48:21.290854 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:21 crc kubenswrapper[4818]: E1122 04:48:21.291026 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.322647 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.322704 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.322723 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.322749 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.322765 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:21Z","lastTransitionTime":"2025-11-22T04:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.425710 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.425761 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.425778 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.425799 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.425816 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:21Z","lastTransitionTime":"2025-11-22T04:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.528415 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.528473 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.528491 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.528513 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.528532 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:21Z","lastTransitionTime":"2025-11-22T04:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.632135 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.632172 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.632183 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.632200 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.632211 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:21Z","lastTransitionTime":"2025-11-22T04:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.681378 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.681456 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.681481 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.681511 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.681535 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:21Z","lastTransitionTime":"2025-11-22T04:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:21 crc kubenswrapper[4818]: E1122 04:48:21.705851 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.710431 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.710507 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.710530 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.710559 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.710581 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:21Z","lastTransitionTime":"2025-11-22T04:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:21 crc kubenswrapper[4818]: E1122 04:48:21.730310 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.733369 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p8s7r_f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/ovnkube-controller/2.log" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.734581 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p8s7r_f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/ovnkube-controller/1.log" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.735805 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.735857 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.735871 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.735888 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.736217 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:21Z","lastTransitionTime":"2025-11-22T04:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.740124 4818 generic.go:334] "Generic (PLEG): container finished" podID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerID="7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b" exitCode=1 Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.740198 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerDied","Data":"7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b"} Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.740349 4818 scope.go:117] "RemoveContainer" containerID="e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.742123 4818 scope.go:117] "RemoveContainer" containerID="7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b" Nov 22 04:48:21 crc kubenswrapper[4818]: E1122 04:48:21.742776 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.759425 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c569fd0932bade53c98da764d0d0e9dd07755d73dbba866c0be0209661aaa05a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: E1122 04:48:21.760954 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.763938 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.763964 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.763973 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.763986 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.763997 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:21Z","lastTransitionTime":"2025-11-22T04:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.770104 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: E1122 04:48:21.775770 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.779131 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.779171 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.779183 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.779200 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.779274 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:21Z","lastTransitionTime":"2025-11-22T04:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.780146 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.789182 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: E1122 04:48:21.791934 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: E1122 04:48:21.792091 4818 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.793653 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.793710 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.793728 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.793774 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.793792 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:21Z","lastTransitionTime":"2025-11-22T04:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.802139 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.816231 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.829315 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.842278 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.853690 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.865840 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.882830 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.896156 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.896475 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.896665 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.896893 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.897096 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:21Z","lastTransitionTime":"2025-11-22T04:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.911461 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668959faaaef590d310c3bb0d59e8739c57d8737861775d86b289a48c4b0832\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"message\\\":\\\"ubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:07.148568 6258 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1122 04:48:07.148772 6258 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:07.149004 6258 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 04:48:07.149078 6258 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 04:48:07.194406 6258 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1122 04:48:07.194444 6258 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1122 04:48:07.194523 6258 ovnkube.go:599] Stopped ovnkube\\\\nI1122 04:48:07.194561 6258 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 04:48:07.194666 6258 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"e column _uuid == {de17f0de-cfb1-4534-bb42-c40f5e050c73}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1122 04:48:21.325074 6483 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}\\\\nI1122 04:48:21.325702 6483 services_controller.go:360] Finished syncing service marketplace-operator-metrics on namespace openshift-marketplace for network=default : 4.744363ms\\\\nI1122 04:48:21.325719 6483 services_controller.go:356] Processing sync for service openshift-kube-scheduler/scheduler for network=default\\\\nI1122 04:48:21.325674 6483 services_controller.go:452] Built service openshift-ingress-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325852 6483 services_controller.go:453] Built service openshift-ingress-operator/metrics template LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325922 6483 services_controller.go:454] Service openshift-ingress-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1122 04:48:21.325012 6483 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.933483 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.950165 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.969208 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:21Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.999672 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.999736 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.999754 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:21 crc kubenswrapper[4818]: I1122 04:48:21.999782 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:21.999806 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:21Z","lastTransitionTime":"2025-11-22T04:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.103580 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.103635 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.103653 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.103675 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.103692 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:22Z","lastTransitionTime":"2025-11-22T04:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.206456 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.206517 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.206533 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.206555 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.206575 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:22Z","lastTransitionTime":"2025-11-22T04:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.290704 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.290802 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:22 crc kubenswrapper[4818]: E1122 04:48:22.290946 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:22 crc kubenswrapper[4818]: E1122 04:48:22.291158 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.292029 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs\") pod \"network-metrics-daemon-djd77\" (UID: \"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\") " pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:22 crc kubenswrapper[4818]: E1122 04:48:22.292228 4818 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 04:48:22 crc kubenswrapper[4818]: E1122 04:48:22.292338 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs podName:429a37e7-cb52-4e76-89ee-16f2b9e6fbfe nodeName:}" failed. No retries permitted until 2025-11-22 04:48:38.292320052 +0000 UTC m=+70.866736579 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs") pod "network-metrics-daemon-djd77" (UID: "429a37e7-cb52-4e76-89ee-16f2b9e6fbfe") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.310445 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.310502 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.310518 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.310542 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.310562 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:22Z","lastTransitionTime":"2025-11-22T04:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.414093 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.414163 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.414183 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.414700 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.414766 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:22Z","lastTransitionTime":"2025-11-22T04:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.517313 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.517398 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.517449 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.517475 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.517492 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:22Z","lastTransitionTime":"2025-11-22T04:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.619365 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.619410 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.619422 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.619439 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.619452 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:22Z","lastTransitionTime":"2025-11-22T04:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.722330 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.722423 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.722452 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.722484 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.722507 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:22Z","lastTransitionTime":"2025-11-22T04:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.747039 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p8s7r_f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/ovnkube-controller/2.log" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.753317 4818 scope.go:117] "RemoveContainer" containerID="7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b" Nov 22 04:48:22 crc kubenswrapper[4818]: E1122 04:48:22.753578 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.769497 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:22Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.786420 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:22Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.806961 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:22Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.826310 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.826389 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.826414 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.826448 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.826468 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:22Z","lastTransitionTime":"2025-11-22T04:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.828309 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:22Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.847290 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:22Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.868024 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:22Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.884360 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:22Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.905094 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:22Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.925661 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:22Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.930461 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.930512 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.930528 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.930550 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.930569 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:22Z","lastTransitionTime":"2025-11-22T04:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.957248 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"e column _uuid == {de17f0de-cfb1-4534-bb42-c40f5e050c73}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1122 04:48:21.325074 6483 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}\\\\nI1122 04:48:21.325702 6483 services_controller.go:360] Finished syncing service marketplace-operator-metrics on namespace openshift-marketplace for network=default : 4.744363ms\\\\nI1122 04:48:21.325719 6483 services_controller.go:356] Processing sync for service openshift-kube-scheduler/scheduler for network=default\\\\nI1122 04:48:21.325674 6483 services_controller.go:452] Built service openshift-ingress-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325852 6483 services_controller.go:453] Built service openshift-ingress-operator/metrics template LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325922 6483 services_controller.go:454] Service openshift-ingress-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1122 04:48:21.325012 6483 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:22Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.977571 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:22Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:22 crc kubenswrapper[4818]: I1122 04:48:22.994848 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:22Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.012776 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.033811 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.033877 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.033894 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.034320 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.034371 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:23Z","lastTransitionTime":"2025-11-22T04:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.035497 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c569fd0932bade53c98da764d0d0e9dd07755d73dbba866c0be0209661aaa05a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.057691 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.137248 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.137330 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.137348 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.137376 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.137394 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:23Z","lastTransitionTime":"2025-11-22T04:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.241153 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.241225 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.241245 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.241325 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.241348 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:23Z","lastTransitionTime":"2025-11-22T04:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.290400 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.290479 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:23 crc kubenswrapper[4818]: E1122 04:48:23.290753 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:23 crc kubenswrapper[4818]: E1122 04:48:23.290929 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.344323 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.344385 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.344402 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.344468 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.344486 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:23Z","lastTransitionTime":"2025-11-22T04:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.447409 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.447460 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.447476 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.447498 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.447515 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:23Z","lastTransitionTime":"2025-11-22T04:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.549877 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.550213 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.550233 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.550289 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.550308 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:23Z","lastTransitionTime":"2025-11-22T04:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.653227 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.653280 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.653289 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.653306 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.653317 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:23Z","lastTransitionTime":"2025-11-22T04:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.665940 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.677659 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.690447 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.707428 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.727606 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.745857 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.756175 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.756241 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.756288 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.756313 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.756331 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:23Z","lastTransitionTime":"2025-11-22T04:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.764367 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.782644 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.803653 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.822955 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.850329 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.854656 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"e column _uuid == {de17f0de-cfb1-4534-bb42-c40f5e050c73}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1122 04:48:21.325074 6483 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}\\\\nI1122 04:48:21.325702 6483 services_controller.go:360] Finished syncing service marketplace-operator-metrics on namespace openshift-marketplace for network=default : 4.744363ms\\\\nI1122 04:48:21.325719 6483 services_controller.go:356] Processing sync for service openshift-kube-scheduler/scheduler for network=default\\\\nI1122 04:48:21.325674 6483 services_controller.go:452] Built service openshift-ingress-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325852 6483 services_controller.go:453] Built service openshift-ingress-operator/metrics template LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325922 6483 services_controller.go:454] Service openshift-ingress-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1122 04:48:21.325012 6483 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.859854 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.859995 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.860178 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.860395 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.860442 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:23Z","lastTransitionTime":"2025-11-22T04:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.864628 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.884774 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.897793 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.914102 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.932873 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c569fd0932bade53c98da764d0d0e9dd07755d73dbba866c0be0209661aaa05a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.949888 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.962724 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.962769 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.962786 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.962807 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.962826 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:23Z","lastTransitionTime":"2025-11-22T04:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.968752 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:23 crc kubenswrapper[4818]: I1122 04:48:23.988025 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.020210 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"e column _uuid == {de17f0de-cfb1-4534-bb42-c40f5e050c73}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1122 04:48:21.325074 6483 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}\\\\nI1122 04:48:21.325702 6483 services_controller.go:360] Finished syncing service marketplace-operator-metrics on namespace openshift-marketplace for network=default : 4.744363ms\\\\nI1122 04:48:21.325719 6483 services_controller.go:356] Processing sync for service openshift-kube-scheduler/scheduler for network=default\\\\nI1122 04:48:21.325674 6483 services_controller.go:452] Built service openshift-ingress-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325852 6483 services_controller.go:453] Built service openshift-ingress-operator/metrics template LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325922 6483 services_controller.go:454] Service openshift-ingress-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1122 04:48:21.325012 6483 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:24Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.038238 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:24Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.053601 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:24Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.065455 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.065492 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.065504 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.065521 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.065532 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:24Z","lastTransitionTime":"2025-11-22T04:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.070011 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:24Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.084985 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:24Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.100548 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c569fd0932bade53c98da764d0d0e9dd07755d73dbba866c0be0209661aaa05a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:24Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.118566 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:24Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.134814 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:24Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.148559 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:24Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.166627 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:24Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.177929 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:24Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.191469 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:24Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.205450 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:24Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.217170 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"affc5026-facf-4e73-a63a-4bfe5ddf7268\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976c49c2e32cd03a00e6a0a91e325864c9a1eb48cc412cc03ed6cdd620da9125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e315f3e5e87e5ed47280e8efc909148c1b18af404778609c867f572a1c352c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe9de7b662c0ba4d848448a8322d3a464e58337895bf25390d4df018b695834a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:24Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.220182 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.220237 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.220269 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.220287 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.220299 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:24Z","lastTransitionTime":"2025-11-22T04:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.290336 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.290407 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:24 crc kubenswrapper[4818]: E1122 04:48:24.290488 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:24 crc kubenswrapper[4818]: E1122 04:48:24.290568 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.322907 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.322955 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.322967 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.322983 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.322996 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:24Z","lastTransitionTime":"2025-11-22T04:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.426010 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.426063 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.426076 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.426094 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.426106 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:24Z","lastTransitionTime":"2025-11-22T04:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.528639 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.528687 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.528700 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.528720 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.528733 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:24Z","lastTransitionTime":"2025-11-22T04:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.631857 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.631915 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.631938 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.631996 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.632031 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:24Z","lastTransitionTime":"2025-11-22T04:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.735425 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.735513 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.735531 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.735564 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.735589 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:24Z","lastTransitionTime":"2025-11-22T04:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.838626 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.838695 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.838770 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.838794 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.838871 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:24Z","lastTransitionTime":"2025-11-22T04:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.942553 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.942658 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.942677 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.942704 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:24 crc kubenswrapper[4818]: I1122 04:48:24.942724 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:24Z","lastTransitionTime":"2025-11-22T04:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.045698 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.045813 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.045830 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.045856 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.045875 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:25Z","lastTransitionTime":"2025-11-22T04:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.151045 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.151112 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.151130 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.151155 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.151173 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:25Z","lastTransitionTime":"2025-11-22T04:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.222411 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:48:25 crc kubenswrapper[4818]: E1122 04:48:25.222658 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:48:57.22261356 +0000 UTC m=+89.797030087 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.222837 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.222920 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:25 crc kubenswrapper[4818]: E1122 04:48:25.223052 4818 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 04:48:25 crc kubenswrapper[4818]: E1122 04:48:25.223089 4818 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 04:48:25 crc kubenswrapper[4818]: E1122 04:48:25.223111 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 04:48:57.223097383 +0000 UTC m=+89.797513910 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 04:48:25 crc kubenswrapper[4818]: E1122 04:48:25.223296 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 04:48:57.223211816 +0000 UTC m=+89.797628353 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.254588 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.254645 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.254659 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.254683 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.254698 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:25Z","lastTransitionTime":"2025-11-22T04:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.291004 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.291127 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:25 crc kubenswrapper[4818]: E1122 04:48:25.291190 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:25 crc kubenswrapper[4818]: E1122 04:48:25.291450 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.324083 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.324189 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:25 crc kubenswrapper[4818]: E1122 04:48:25.324416 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 04:48:25 crc kubenswrapper[4818]: E1122 04:48:25.324464 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 04:48:25 crc kubenswrapper[4818]: E1122 04:48:25.324486 4818 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:48:25 crc kubenswrapper[4818]: E1122 04:48:25.324522 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 04:48:25 crc kubenswrapper[4818]: E1122 04:48:25.324579 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 04:48:57.324553848 +0000 UTC m=+89.898970405 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:48:25 crc kubenswrapper[4818]: E1122 04:48:25.324579 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 04:48:25 crc kubenswrapper[4818]: E1122 04:48:25.324612 4818 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:48:25 crc kubenswrapper[4818]: E1122 04:48:25.324690 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 04:48:57.324665131 +0000 UTC m=+89.899081688 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.358668 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.358755 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.358785 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.358817 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.358882 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:25Z","lastTransitionTime":"2025-11-22T04:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.462015 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.462066 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.462084 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.462108 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.462125 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:25Z","lastTransitionTime":"2025-11-22T04:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.565892 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.565965 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.565983 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.566006 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.566025 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:25Z","lastTransitionTime":"2025-11-22T04:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.669529 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.669612 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.669634 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.669659 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.669677 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:25Z","lastTransitionTime":"2025-11-22T04:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.772202 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.772306 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.772331 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.772364 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.772387 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:25Z","lastTransitionTime":"2025-11-22T04:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.875859 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.875920 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.875936 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.875963 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.875982 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:25Z","lastTransitionTime":"2025-11-22T04:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.979299 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.979366 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.979387 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.979424 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:25 crc kubenswrapper[4818]: I1122 04:48:25.979446 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:25Z","lastTransitionTime":"2025-11-22T04:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.083066 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.083138 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.083158 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.083183 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.083200 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:26Z","lastTransitionTime":"2025-11-22T04:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.187333 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.187398 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.187418 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.187441 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.187459 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:26Z","lastTransitionTime":"2025-11-22T04:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.290475 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.290512 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.290519 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.290552 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.290651 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.290690 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:26 crc kubenswrapper[4818]: E1122 04:48:26.290699 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.290720 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:26Z","lastTransitionTime":"2025-11-22T04:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:26 crc kubenswrapper[4818]: E1122 04:48:26.290809 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.394200 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.394314 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.394340 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.394373 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.394399 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:26Z","lastTransitionTime":"2025-11-22T04:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.497436 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.497526 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.497557 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.497587 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.497609 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:26Z","lastTransitionTime":"2025-11-22T04:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.601226 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.601340 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.601364 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.601397 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.601420 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:26Z","lastTransitionTime":"2025-11-22T04:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.704247 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.704340 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.704359 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.704386 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.704406 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:26Z","lastTransitionTime":"2025-11-22T04:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.807129 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.807195 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.807216 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.807240 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.807287 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:26Z","lastTransitionTime":"2025-11-22T04:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.910958 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.911031 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.911048 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.911073 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:26 crc kubenswrapper[4818]: I1122 04:48:26.911090 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:26Z","lastTransitionTime":"2025-11-22T04:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.014120 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.014179 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.014198 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.014219 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.014236 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:27Z","lastTransitionTime":"2025-11-22T04:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.117432 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.117496 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.117516 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.117540 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.117558 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:27Z","lastTransitionTime":"2025-11-22T04:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.220492 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.220551 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.220567 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.220589 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.220607 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:27Z","lastTransitionTime":"2025-11-22T04:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.290926 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:27 crc kubenswrapper[4818]: E1122 04:48:27.291093 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.291669 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:27 crc kubenswrapper[4818]: E1122 04:48:27.291770 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.323976 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.324047 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.324065 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.324089 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.324109 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:27Z","lastTransitionTime":"2025-11-22T04:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.426664 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.426706 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.426715 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.426729 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.426738 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:27Z","lastTransitionTime":"2025-11-22T04:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.530562 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.530627 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.530644 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.530673 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.530691 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:27Z","lastTransitionTime":"2025-11-22T04:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.634156 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.634221 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.634238 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.634298 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.634316 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:27Z","lastTransitionTime":"2025-11-22T04:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.737538 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.737606 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.737630 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.737659 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.737681 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:27Z","lastTransitionTime":"2025-11-22T04:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.840559 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.840632 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.840666 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.840695 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.840719 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:27Z","lastTransitionTime":"2025-11-22T04:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.944019 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.944105 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.944123 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.944152 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:27 crc kubenswrapper[4818]: I1122 04:48:27.944170 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:27Z","lastTransitionTime":"2025-11-22T04:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.047524 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.047597 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.047614 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.047638 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.047656 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:28Z","lastTransitionTime":"2025-11-22T04:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.150945 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.150982 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.150990 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.151008 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.151021 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:28Z","lastTransitionTime":"2025-11-22T04:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.253434 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.253482 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.253500 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.253522 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.253538 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:28Z","lastTransitionTime":"2025-11-22T04:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.290281 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.290354 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:28 crc kubenswrapper[4818]: E1122 04:48:28.290528 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:28 crc kubenswrapper[4818]: E1122 04:48:28.290667 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.311224 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:28Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.329601 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:28Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.351073 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c569fd0932bade53c98da764d0d0e9dd07755d73dbba866c0be0209661aaa05a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:28Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.356655 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.356696 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.356710 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.356730 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.356746 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:28Z","lastTransitionTime":"2025-11-22T04:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.368552 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:28Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.388845 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:28Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.408872 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:28Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.426598 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:28Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.437986 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:28Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.448748 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:28Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.458988 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.459033 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.459047 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.459065 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.459077 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:28Z","lastTransitionTime":"2025-11-22T04:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.460445 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"affc5026-facf-4e73-a63a-4bfe5ddf7268\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976c49c2e32cd03a00e6a0a91e325864c9a1eb48cc412cc03ed6cdd620da9125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e315f3e5e87e5ed47280e8efc909148c1b18af404778609c867f572a1c352c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe9de7b662c0ba4d848448a8322d3a464e58337895bf25390d4df018b695834a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:28Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.470986 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:28Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.482309 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:28Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.500933 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"e column _uuid == {de17f0de-cfb1-4534-bb42-c40f5e050c73}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1122 04:48:21.325074 6483 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}\\\\nI1122 04:48:21.325702 6483 services_controller.go:360] Finished syncing service marketplace-operator-metrics on namespace openshift-marketplace for network=default : 4.744363ms\\\\nI1122 04:48:21.325719 6483 services_controller.go:356] Processing sync for service openshift-kube-scheduler/scheduler for network=default\\\\nI1122 04:48:21.325674 6483 services_controller.go:452] Built service openshift-ingress-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325852 6483 services_controller.go:453] Built service openshift-ingress-operator/metrics template LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325922 6483 services_controller.go:454] Service openshift-ingress-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1122 04:48:21.325012 6483 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:28Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.552236 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:28Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.561780 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.561833 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.561844 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.561859 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.561869 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:28Z","lastTransitionTime":"2025-11-22T04:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.564180 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:28Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.577789 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:28Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.664049 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.664102 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.664110 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.664122 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.664131 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:28Z","lastTransitionTime":"2025-11-22T04:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.766133 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.766480 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.766494 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.766511 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.766524 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:28Z","lastTransitionTime":"2025-11-22T04:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.868470 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.868505 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.868513 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.868527 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.868536 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:28Z","lastTransitionTime":"2025-11-22T04:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.970270 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.970324 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.970334 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.970346 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:28 crc kubenswrapper[4818]: I1122 04:48:28.970354 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:28Z","lastTransitionTime":"2025-11-22T04:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.072720 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.072778 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.072802 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.072829 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.072850 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:29Z","lastTransitionTime":"2025-11-22T04:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.175513 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.175560 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.175575 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.175597 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.175614 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:29Z","lastTransitionTime":"2025-11-22T04:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.277732 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.277905 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.278008 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.278111 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.278202 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:29Z","lastTransitionTime":"2025-11-22T04:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.290053 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.290061 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:29 crc kubenswrapper[4818]: E1122 04:48:29.290206 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:29 crc kubenswrapper[4818]: E1122 04:48:29.290431 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.380298 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.380351 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.380366 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.380387 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.380405 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:29Z","lastTransitionTime":"2025-11-22T04:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.482767 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.482804 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.482815 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.482830 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.482842 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:29Z","lastTransitionTime":"2025-11-22T04:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.585098 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.585163 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.585183 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.585205 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.585219 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:29Z","lastTransitionTime":"2025-11-22T04:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.687441 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.687478 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.687488 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.687504 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.687514 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:29Z","lastTransitionTime":"2025-11-22T04:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.789790 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.789832 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.789843 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.789859 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.789870 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:29Z","lastTransitionTime":"2025-11-22T04:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.892150 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.892189 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.892199 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.892225 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.892237 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:29Z","lastTransitionTime":"2025-11-22T04:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.995027 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.995099 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.995109 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.995122 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:29 crc kubenswrapper[4818]: I1122 04:48:29.995131 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:29Z","lastTransitionTime":"2025-11-22T04:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.098023 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.098060 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.098075 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.098094 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.098104 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:30Z","lastTransitionTime":"2025-11-22T04:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.199930 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.199996 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.200008 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.200021 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.200031 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:30Z","lastTransitionTime":"2025-11-22T04:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.290441 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:30 crc kubenswrapper[4818]: E1122 04:48:30.290814 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.290465 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:30 crc kubenswrapper[4818]: E1122 04:48:30.291120 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.301928 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.302141 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.302273 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.302396 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.302442 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:30Z","lastTransitionTime":"2025-11-22T04:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.405043 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.405127 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.405170 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.405188 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.405198 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:30Z","lastTransitionTime":"2025-11-22T04:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.508282 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.508367 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.508385 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.508408 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.508424 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:30Z","lastTransitionTime":"2025-11-22T04:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.611520 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.611584 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.611595 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.611611 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.611623 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:30Z","lastTransitionTime":"2025-11-22T04:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.714678 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.714740 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.714761 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.714789 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.714812 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:30Z","lastTransitionTime":"2025-11-22T04:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.817202 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.817245 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.817280 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.817296 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.817307 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:30Z","lastTransitionTime":"2025-11-22T04:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.923332 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.923835 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.923868 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.923897 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:30 crc kubenswrapper[4818]: I1122 04:48:30.923916 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:30Z","lastTransitionTime":"2025-11-22T04:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.027806 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.027881 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.027905 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.027934 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.027955 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:31Z","lastTransitionTime":"2025-11-22T04:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.136035 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.136120 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.136133 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.136161 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.136178 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:31Z","lastTransitionTime":"2025-11-22T04:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.238982 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.239044 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.239062 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.239083 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.239101 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:31Z","lastTransitionTime":"2025-11-22T04:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.291016 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.291143 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:31 crc kubenswrapper[4818]: E1122 04:48:31.291272 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:31 crc kubenswrapper[4818]: E1122 04:48:31.291420 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.342111 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.342150 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.342160 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.342173 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.342184 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:31Z","lastTransitionTime":"2025-11-22T04:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.445972 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.446019 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.446039 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.446061 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.446078 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:31Z","lastTransitionTime":"2025-11-22T04:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.555924 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.555965 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.555976 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.555991 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.556000 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:31Z","lastTransitionTime":"2025-11-22T04:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.658160 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.658201 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.658211 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.658225 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.658234 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:31Z","lastTransitionTime":"2025-11-22T04:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.761474 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.761531 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.761548 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.761572 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.761589 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:31Z","lastTransitionTime":"2025-11-22T04:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.824340 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.824384 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.824401 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.824422 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.824439 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:31Z","lastTransitionTime":"2025-11-22T04:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:31 crc kubenswrapper[4818]: E1122 04:48:31.845635 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:31Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.851180 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.851296 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.851325 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.851357 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.851380 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:31Z","lastTransitionTime":"2025-11-22T04:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:31 crc kubenswrapper[4818]: E1122 04:48:31.867460 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:31Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.872369 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.872440 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.872465 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.872495 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.872521 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:31Z","lastTransitionTime":"2025-11-22T04:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:31 crc kubenswrapper[4818]: E1122 04:48:31.906794 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:31Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.913162 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.913215 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.913233 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.913261 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.913296 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:31Z","lastTransitionTime":"2025-11-22T04:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:31 crc kubenswrapper[4818]: E1122 04:48:31.931212 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:31Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.938551 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.938602 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.938613 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.938627 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.938636 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:31Z","lastTransitionTime":"2025-11-22T04:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:31 crc kubenswrapper[4818]: E1122 04:48:31.952880 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:31Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:31 crc kubenswrapper[4818]: E1122 04:48:31.953006 4818 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.954452 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.954517 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.954535 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.954555 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:31 crc kubenswrapper[4818]: I1122 04:48:31.954573 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:31Z","lastTransitionTime":"2025-11-22T04:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.057593 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.057828 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.057891 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.057950 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.058030 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:32Z","lastTransitionTime":"2025-11-22T04:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.161320 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.161399 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.161423 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.161456 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.161480 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:32Z","lastTransitionTime":"2025-11-22T04:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.267180 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.267296 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.267324 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.267360 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.267596 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:32Z","lastTransitionTime":"2025-11-22T04:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.291322 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.291353 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:32 crc kubenswrapper[4818]: E1122 04:48:32.292147 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:32 crc kubenswrapper[4818]: E1122 04:48:32.292057 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.370306 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.370366 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.370383 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.370406 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.370423 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:32Z","lastTransitionTime":"2025-11-22T04:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.473560 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.473621 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.473638 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.473665 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.473683 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:32Z","lastTransitionTime":"2025-11-22T04:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.576550 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.576689 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.576709 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.576732 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.576749 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:32Z","lastTransitionTime":"2025-11-22T04:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.679740 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.679803 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.679823 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.679847 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.679865 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:32Z","lastTransitionTime":"2025-11-22T04:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.783307 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.783382 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.783394 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.783412 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.783423 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:32Z","lastTransitionTime":"2025-11-22T04:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.885763 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.885836 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.885862 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.885895 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.885919 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:32Z","lastTransitionTime":"2025-11-22T04:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.989413 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.989473 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.989495 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.989516 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:32 crc kubenswrapper[4818]: I1122 04:48:32.989533 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:32Z","lastTransitionTime":"2025-11-22T04:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.093178 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.093235 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.093291 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.093323 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.093375 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:33Z","lastTransitionTime":"2025-11-22T04:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.196582 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.196680 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.196705 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.196733 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.196750 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:33Z","lastTransitionTime":"2025-11-22T04:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.290744 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.290753 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:33 crc kubenswrapper[4818]: E1122 04:48:33.290926 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:33 crc kubenswrapper[4818]: E1122 04:48:33.290975 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.299203 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.299235 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.299246 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.299278 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.299293 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:33Z","lastTransitionTime":"2025-11-22T04:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.402305 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.402347 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.402355 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.402368 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.402377 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:33Z","lastTransitionTime":"2025-11-22T04:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.504109 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.504141 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.504149 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.504163 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.504172 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:33Z","lastTransitionTime":"2025-11-22T04:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.606329 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.606567 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.606632 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.606706 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.606780 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:33Z","lastTransitionTime":"2025-11-22T04:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.709434 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.709492 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.709510 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.709532 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.709551 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:33Z","lastTransitionTime":"2025-11-22T04:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.811806 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.812054 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.812241 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.812569 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.812741 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:33Z","lastTransitionTime":"2025-11-22T04:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.915283 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.915318 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.915327 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.915340 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:33 crc kubenswrapper[4818]: I1122 04:48:33.915351 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:33Z","lastTransitionTime":"2025-11-22T04:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.018647 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.018690 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.018698 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.018712 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.018728 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:34Z","lastTransitionTime":"2025-11-22T04:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.121415 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.121450 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.121459 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.121471 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.121479 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:34Z","lastTransitionTime":"2025-11-22T04:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.224148 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.224175 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.224184 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.224196 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.224206 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:34Z","lastTransitionTime":"2025-11-22T04:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.290344 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.290398 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:34 crc kubenswrapper[4818]: E1122 04:48:34.290485 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:34 crc kubenswrapper[4818]: E1122 04:48:34.290599 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.291300 4818 scope.go:117] "RemoveContainer" containerID="7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b" Nov 22 04:48:34 crc kubenswrapper[4818]: E1122 04:48:34.291456 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.326824 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.326868 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.326880 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.326898 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.326910 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:34Z","lastTransitionTime":"2025-11-22T04:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.429235 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.429521 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.429614 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.429692 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.429761 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:34Z","lastTransitionTime":"2025-11-22T04:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.532565 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.532593 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.532601 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.532614 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.532623 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:34Z","lastTransitionTime":"2025-11-22T04:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.635002 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.635066 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.635089 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.635121 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.635160 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:34Z","lastTransitionTime":"2025-11-22T04:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.737835 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.737890 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.737915 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.737943 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.737965 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:34Z","lastTransitionTime":"2025-11-22T04:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.840331 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.840385 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.840402 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.840424 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.840440 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:34Z","lastTransitionTime":"2025-11-22T04:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.942131 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.942158 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.942167 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.942179 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:34 crc kubenswrapper[4818]: I1122 04:48:34.942187 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:34Z","lastTransitionTime":"2025-11-22T04:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.045605 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.045657 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.045673 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.045695 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.045714 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:35Z","lastTransitionTime":"2025-11-22T04:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.148148 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.148559 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.148789 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.149049 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.149315 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:35Z","lastTransitionTime":"2025-11-22T04:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.252545 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.252619 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.252637 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.253032 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.253085 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:35Z","lastTransitionTime":"2025-11-22T04:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.290742 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.290742 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:35 crc kubenswrapper[4818]: E1122 04:48:35.290934 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:35 crc kubenswrapper[4818]: E1122 04:48:35.291072 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.356117 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.356184 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.356209 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.356236 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.356321 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:35Z","lastTransitionTime":"2025-11-22T04:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.458314 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.458535 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.458631 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.458746 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.458833 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:35Z","lastTransitionTime":"2025-11-22T04:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.561839 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.561896 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.561918 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.561945 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.561967 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:35Z","lastTransitionTime":"2025-11-22T04:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.664548 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.664584 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.664594 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.664611 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.664621 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:35Z","lastTransitionTime":"2025-11-22T04:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.767144 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.767437 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.767514 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.767590 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.767654 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:35Z","lastTransitionTime":"2025-11-22T04:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.870500 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.870570 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.870594 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.870618 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.870636 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:35Z","lastTransitionTime":"2025-11-22T04:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.972672 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.972824 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.972904 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.972986 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:35 crc kubenswrapper[4818]: I1122 04:48:35.973051 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:35Z","lastTransitionTime":"2025-11-22T04:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.075426 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.075665 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.075777 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.075870 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.075965 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:36Z","lastTransitionTime":"2025-11-22T04:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.178657 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.178697 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.178706 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.178719 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.178728 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:36Z","lastTransitionTime":"2025-11-22T04:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.281552 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.281621 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.281644 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.281666 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.281684 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:36Z","lastTransitionTime":"2025-11-22T04:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.290279 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.290433 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:36 crc kubenswrapper[4818]: E1122 04:48:36.290662 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:36 crc kubenswrapper[4818]: E1122 04:48:36.290769 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.384468 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.384527 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.384544 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.384569 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.384585 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:36Z","lastTransitionTime":"2025-11-22T04:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.486770 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.487250 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.487492 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.487695 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.487891 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:36Z","lastTransitionTime":"2025-11-22T04:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.590924 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.591433 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.591504 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.591586 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.591658 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:36Z","lastTransitionTime":"2025-11-22T04:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.723344 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.723384 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.723396 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.723418 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.723433 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:36Z","lastTransitionTime":"2025-11-22T04:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.825512 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.825558 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.825572 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.825606 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.825618 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:36Z","lastTransitionTime":"2025-11-22T04:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.928159 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.928468 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.928535 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.928617 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:36 crc kubenswrapper[4818]: I1122 04:48:36.928687 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:36Z","lastTransitionTime":"2025-11-22T04:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.030496 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.030752 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.030831 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.030900 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.030956 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:37Z","lastTransitionTime":"2025-11-22T04:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.133608 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.133659 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.133674 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.133693 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.133704 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:37Z","lastTransitionTime":"2025-11-22T04:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.236423 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.236473 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.236484 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.236497 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.236507 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:37Z","lastTransitionTime":"2025-11-22T04:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.290986 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:37 crc kubenswrapper[4818]: E1122 04:48:37.291103 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.291279 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:37 crc kubenswrapper[4818]: E1122 04:48:37.291469 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.339076 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.339345 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.339451 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.339552 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.339649 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:37Z","lastTransitionTime":"2025-11-22T04:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.441158 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.441480 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.441556 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.441649 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.441714 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:37Z","lastTransitionTime":"2025-11-22T04:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.544305 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.544368 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.544404 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.544434 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.544456 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:37Z","lastTransitionTime":"2025-11-22T04:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.647157 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.647193 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.647202 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.647215 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.647224 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:37Z","lastTransitionTime":"2025-11-22T04:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.748838 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.748874 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.748883 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.748897 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.748906 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:37Z","lastTransitionTime":"2025-11-22T04:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.853774 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.853826 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.853835 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.853848 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.853858 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:37Z","lastTransitionTime":"2025-11-22T04:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.956246 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.956309 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.956321 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.956338 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:37 crc kubenswrapper[4818]: I1122 04:48:37.956349 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:37Z","lastTransitionTime":"2025-11-22T04:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.058873 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.058939 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.058956 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.058980 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.058997 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:38Z","lastTransitionTime":"2025-11-22T04:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.161700 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.161739 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.161750 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.161766 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.161776 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:38Z","lastTransitionTime":"2025-11-22T04:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.263899 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.263952 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.263963 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.263977 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.263989 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:38Z","lastTransitionTime":"2025-11-22T04:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.290486 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.290566 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:38 crc kubenswrapper[4818]: E1122 04:48:38.290595 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:38 crc kubenswrapper[4818]: E1122 04:48:38.290758 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.311844 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:38Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.330525 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:38Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.359507 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"e column _uuid == {de17f0de-cfb1-4534-bb42-c40f5e050c73}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1122 04:48:21.325074 6483 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}\\\\nI1122 04:48:21.325702 6483 services_controller.go:360] Finished syncing service marketplace-operator-metrics on namespace openshift-marketplace for network=default : 4.744363ms\\\\nI1122 04:48:21.325719 6483 services_controller.go:356] Processing sync for service openshift-kube-scheduler/scheduler for network=default\\\\nI1122 04:48:21.325674 6483 services_controller.go:452] Built service openshift-ingress-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325852 6483 services_controller.go:453] Built service openshift-ingress-operator/metrics template LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325922 6483 services_controller.go:454] Service openshift-ingress-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1122 04:48:21.325012 6483 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:38Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.363296 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs\") pod \"network-metrics-daemon-djd77\" (UID: \"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\") " pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:38 crc kubenswrapper[4818]: E1122 04:48:38.363543 4818 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 04:48:38 crc kubenswrapper[4818]: E1122 04:48:38.363645 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs podName:429a37e7-cb52-4e76-89ee-16f2b9e6fbfe nodeName:}" failed. No retries permitted until 2025-11-22 04:49:10.363615185 +0000 UTC m=+102.938031752 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs") pod "network-metrics-daemon-djd77" (UID: "429a37e7-cb52-4e76-89ee-16f2b9e6fbfe") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.367025 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.367092 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.367104 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.367116 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.367126 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:38Z","lastTransitionTime":"2025-11-22T04:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.373522 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:38Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.383005 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:38Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.394143 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:38Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.406176 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c569fd0932bade53c98da764d0d0e9dd07755d73dbba866c0be0209661aaa05a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:38Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.418156 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:38Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.428012 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:38Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.441312 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"affc5026-facf-4e73-a63a-4bfe5ddf7268\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976c49c2e32cd03a00e6a0a91e325864c9a1eb48cc412cc03ed6cdd620da9125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e315f3e5e87e5ed47280e8efc909148c1b18af404778609c867f572a1c352c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe9de7b662c0ba4d848448a8322d3a464e58337895bf25390d4df018b695834a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:38Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.459735 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:38Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.469597 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.469626 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.469638 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.469653 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.469732 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:38Z","lastTransitionTime":"2025-11-22T04:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.476334 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:38Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.493827 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:38Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.511536 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:38Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.525008 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:38Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.538158 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:38Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.571387 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.571432 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.571443 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.571460 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.571471 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:38Z","lastTransitionTime":"2025-11-22T04:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.674159 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.674221 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.674231 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.674587 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.674644 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:38Z","lastTransitionTime":"2025-11-22T04:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.777173 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.777219 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.777230 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.777266 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.777279 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:38Z","lastTransitionTime":"2025-11-22T04:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.880147 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.880204 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.880220 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.880240 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.880272 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:38Z","lastTransitionTime":"2025-11-22T04:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.982525 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.982583 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.982596 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.982616 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:38 crc kubenswrapper[4818]: I1122 04:48:38.982630 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:38Z","lastTransitionTime":"2025-11-22T04:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.084727 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.084785 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.084803 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.084834 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.084856 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:39Z","lastTransitionTime":"2025-11-22T04:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.188066 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.188129 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.188147 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.188173 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.188191 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:39Z","lastTransitionTime":"2025-11-22T04:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.290066 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:39 crc kubenswrapper[4818]: E1122 04:48:39.290218 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.290339 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:39 crc kubenswrapper[4818]: E1122 04:48:39.290419 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.290910 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.290944 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.290953 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.290966 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.290975 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:39Z","lastTransitionTime":"2025-11-22T04:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.394192 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.394283 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.394302 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.394326 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.394343 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:39Z","lastTransitionTime":"2025-11-22T04:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.497816 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.497882 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.497908 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.497938 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.497964 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:39Z","lastTransitionTime":"2025-11-22T04:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.600845 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.600904 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.600922 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.600960 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.600978 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:39Z","lastTransitionTime":"2025-11-22T04:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.703747 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.703800 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.703818 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.703840 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.703857 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:39Z","lastTransitionTime":"2025-11-22T04:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.807658 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.807726 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.807748 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.807778 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.807800 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:39Z","lastTransitionTime":"2025-11-22T04:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.911373 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.911443 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.911466 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.911497 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:39 crc kubenswrapper[4818]: I1122 04:48:39.911524 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:39Z","lastTransitionTime":"2025-11-22T04:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.015005 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.015062 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.015080 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.015109 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.015127 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:40Z","lastTransitionTime":"2025-11-22T04:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.117438 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.117489 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.117507 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.117529 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.117544 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:40Z","lastTransitionTime":"2025-11-22T04:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.220416 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.220483 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.220500 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.220524 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.220541 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:40Z","lastTransitionTime":"2025-11-22T04:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.290791 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:40 crc kubenswrapper[4818]: E1122 04:48:40.290982 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.291648 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:40 crc kubenswrapper[4818]: E1122 04:48:40.292004 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.322762 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.322797 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.322810 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.322827 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.322840 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:40Z","lastTransitionTime":"2025-11-22T04:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.425585 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.425981 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.426145 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.426335 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.426510 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:40Z","lastTransitionTime":"2025-11-22T04:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.529537 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.529595 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.529611 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.529635 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.529652 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:40Z","lastTransitionTime":"2025-11-22T04:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.632192 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.632287 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.632305 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.632329 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.632353 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:40Z","lastTransitionTime":"2025-11-22T04:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.735814 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.736101 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.736174 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.736355 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.736489 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:40Z","lastTransitionTime":"2025-11-22T04:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.839292 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.839334 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.839345 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.839363 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.839375 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:40Z","lastTransitionTime":"2025-11-22T04:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.942053 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.942124 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.942141 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.942166 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:40 crc kubenswrapper[4818]: I1122 04:48:40.942185 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:40Z","lastTransitionTime":"2025-11-22T04:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.044756 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.044803 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.044814 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.044829 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.044840 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:41Z","lastTransitionTime":"2025-11-22T04:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.147141 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.147171 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.147182 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.147197 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.147208 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:41Z","lastTransitionTime":"2025-11-22T04:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.250083 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.250115 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.250126 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.250140 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.250150 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:41Z","lastTransitionTime":"2025-11-22T04:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.290597 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.290685 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:41 crc kubenswrapper[4818]: E1122 04:48:41.290877 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:41 crc kubenswrapper[4818]: E1122 04:48:41.291090 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.355715 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.355955 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.356041 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.356133 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.356209 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:41Z","lastTransitionTime":"2025-11-22T04:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.459197 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.459247 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.459276 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.459295 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.459308 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:41Z","lastTransitionTime":"2025-11-22T04:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.561438 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.561482 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.561493 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.561513 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.561528 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:41Z","lastTransitionTime":"2025-11-22T04:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.664195 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.664227 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.664235 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.664267 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.664280 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:41Z","lastTransitionTime":"2025-11-22T04:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.766688 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.766733 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.766742 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.766757 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.766770 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:41Z","lastTransitionTime":"2025-11-22T04:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.869555 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.869618 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.869640 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.869672 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.869694 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:41Z","lastTransitionTime":"2025-11-22T04:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.972728 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.972810 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.972835 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.972867 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:41 crc kubenswrapper[4818]: I1122 04:48:41.972890 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:41Z","lastTransitionTime":"2025-11-22T04:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.038506 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.038567 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.038578 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.038598 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.038611 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:42Z","lastTransitionTime":"2025-11-22T04:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:42 crc kubenswrapper[4818]: E1122 04:48:42.060622 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:42Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.066444 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.066507 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.066527 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.066556 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.066580 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:42Z","lastTransitionTime":"2025-11-22T04:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:42 crc kubenswrapper[4818]: E1122 04:48:42.081831 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:42Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.086046 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.086099 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.086117 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.086140 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.086156 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:42Z","lastTransitionTime":"2025-11-22T04:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:42 crc kubenswrapper[4818]: E1122 04:48:42.108954 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:42Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.112847 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.112899 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.112915 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.112937 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.112955 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:42Z","lastTransitionTime":"2025-11-22T04:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:42 crc kubenswrapper[4818]: E1122 04:48:42.128883 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:42Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.133082 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.133129 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.133146 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.133169 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.133185 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:42Z","lastTransitionTime":"2025-11-22T04:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:42 crc kubenswrapper[4818]: E1122 04:48:42.151794 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:42Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:42 crc kubenswrapper[4818]: E1122 04:48:42.151979 4818 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.154602 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.154655 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.154678 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.154700 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.154718 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:42Z","lastTransitionTime":"2025-11-22T04:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.257922 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.257997 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.258018 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.258040 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.258061 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:42Z","lastTransitionTime":"2025-11-22T04:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.290635 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.290688 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:42 crc kubenswrapper[4818]: E1122 04:48:42.290828 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:42 crc kubenswrapper[4818]: E1122 04:48:42.290959 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.359878 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.359958 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.359981 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.360008 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.360028 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:42Z","lastTransitionTime":"2025-11-22T04:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.462673 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.462722 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.462738 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.462763 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.462780 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:42Z","lastTransitionTime":"2025-11-22T04:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.565314 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.565375 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.565391 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.565415 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.565434 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:42Z","lastTransitionTime":"2025-11-22T04:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.668356 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.668461 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.668484 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.668512 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.668531 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:42Z","lastTransitionTime":"2025-11-22T04:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.771993 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.772035 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.772044 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.772060 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.772072 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:42Z","lastTransitionTime":"2025-11-22T04:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.820778 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x865d_d96879f8-2766-4f5e-bc3e-bbf8e5394a94/kube-multus/0.log" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.820863 4818 generic.go:334] "Generic (PLEG): container finished" podID="d96879f8-2766-4f5e-bc3e-bbf8e5394a94" containerID="27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2" exitCode=1 Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.820932 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x865d" event={"ID":"d96879f8-2766-4f5e-bc3e-bbf8e5394a94","Type":"ContainerDied","Data":"27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2"} Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.822066 4818 scope.go:117] "RemoveContainer" containerID="27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.842441 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c569fd0932bade53c98da764d0d0e9dd07755d73dbba866c0be0209661aaa05a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:42Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.862319 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"2025-11-22T04:47:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e6acc9f9-59f7-401f-8660-bd7d0fc9e096\\\\n2025-11-22T04:47:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e6acc9f9-59f7-401f-8660-bd7d0fc9e096 to /host/opt/cni/bin/\\\\n2025-11-22T04:47:57Z [verbose] multus-daemon started\\\\n2025-11-22T04:47:57Z [verbose] Readiness Indicator file check\\\\n2025-11-22T04:48:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:42Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.874644 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.874764 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.874849 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.874936 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.875016 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:42Z","lastTransitionTime":"2025-11-22T04:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.883434 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:42Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.898007 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:42Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.910459 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:42Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.926668 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:42Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.940117 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"affc5026-facf-4e73-a63a-4bfe5ddf7268\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976c49c2e32cd03a00e6a0a91e325864c9a1eb48cc412cc03ed6cdd620da9125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e315f3e5e87e5ed47280e8efc909148c1b18af404778609c867f572a1c352c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe9de7b662c0ba4d848448a8322d3a464e58337895bf25390d4df018b695834a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:42Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.953711 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:42Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.975471 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:42Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.977384 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.977583 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.977729 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.977872 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.978010 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:42Z","lastTransitionTime":"2025-11-22T04:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:42 crc kubenswrapper[4818]: I1122 04:48:42.996661 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:42Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.020143 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:43Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.033003 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:43Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.049881 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:43Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.071516 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"e column _uuid == {de17f0de-cfb1-4534-bb42-c40f5e050c73}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1122 04:48:21.325074 6483 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}\\\\nI1122 04:48:21.325702 6483 services_controller.go:360] Finished syncing service marketplace-operator-metrics on namespace openshift-marketplace for network=default : 4.744363ms\\\\nI1122 04:48:21.325719 6483 services_controller.go:356] Processing sync for service openshift-kube-scheduler/scheduler for network=default\\\\nI1122 04:48:21.325674 6483 services_controller.go:452] Built service openshift-ingress-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325852 6483 services_controller.go:453] Built service openshift-ingress-operator/metrics template LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325922 6483 services_controller.go:454] Service openshift-ingress-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1122 04:48:21.325012 6483 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:43Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.081101 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.081130 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.081140 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.081155 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.081168 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:43Z","lastTransitionTime":"2025-11-22T04:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.085599 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:43Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.101072 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:43Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.183579 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.183631 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.183650 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.183673 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.183701 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:43Z","lastTransitionTime":"2025-11-22T04:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.286539 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.286584 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.286597 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.286612 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.286624 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:43Z","lastTransitionTime":"2025-11-22T04:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.290984 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.291082 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:43 crc kubenswrapper[4818]: E1122 04:48:43.291101 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:43 crc kubenswrapper[4818]: E1122 04:48:43.291275 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.389771 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.389815 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.389824 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.389841 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.389851 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:43Z","lastTransitionTime":"2025-11-22T04:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.492408 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.492482 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.492492 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.492506 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.492515 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:43Z","lastTransitionTime":"2025-11-22T04:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.596772 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.596827 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.596845 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.596868 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.596886 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:43Z","lastTransitionTime":"2025-11-22T04:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.699909 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.699955 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.699973 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.699996 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.700014 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:43Z","lastTransitionTime":"2025-11-22T04:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.803360 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.803402 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.803414 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.803429 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.803441 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:43Z","lastTransitionTime":"2025-11-22T04:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.827144 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x865d_d96879f8-2766-4f5e-bc3e-bbf8e5394a94/kube-multus/0.log" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.827235 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x865d" event={"ID":"d96879f8-2766-4f5e-bc3e-bbf8e5394a94","Type":"ContainerStarted","Data":"52d352c1148218337557f43c9772c5cc9866398ef20d8b5eb91eeff53c4cecbd"} Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.850468 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c569fd0932bade53c98da764d0d0e9dd07755d73dbba866c0be0209661aaa05a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:43Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.871445 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52d352c1148218337557f43c9772c5cc9866398ef20d8b5eb91eeff53c4cecbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"2025-11-22T04:47:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e6acc9f9-59f7-401f-8660-bd7d0fc9e096\\\\n2025-11-22T04:47:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e6acc9f9-59f7-401f-8660-bd7d0fc9e096 to /host/opt/cni/bin/\\\\n2025-11-22T04:47:57Z [verbose] multus-daemon started\\\\n2025-11-22T04:47:57Z [verbose] Readiness Indicator file check\\\\n2025-11-22T04:48:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:43Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.890894 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:43Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.906741 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.906787 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.906806 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.906834 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.906860 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:43Z","lastTransitionTime":"2025-11-22T04:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.906966 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:43Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.924034 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:43Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.944235 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"affc5026-facf-4e73-a63a-4bfe5ddf7268\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976c49c2e32cd03a00e6a0a91e325864c9a1eb48cc412cc03ed6cdd620da9125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e315f3e5e87e5ed47280e8efc909148c1b18af404778609c867f572a1c352c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe9de7b662c0ba4d848448a8322d3a464e58337895bf25390d4df018b695834a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:43Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.966091 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:43Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:43 crc kubenswrapper[4818]: I1122 04:48:43.989084 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:43Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.007116 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:44Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.010379 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.010414 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.010600 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.010631 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.010645 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:44Z","lastTransitionTime":"2025-11-22T04:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.026820 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:44Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.045492 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:44Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.068716 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:44Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.097318 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"e column _uuid == {de17f0de-cfb1-4534-bb42-c40f5e050c73}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1122 04:48:21.325074 6483 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}\\\\nI1122 04:48:21.325702 6483 services_controller.go:360] Finished syncing service marketplace-operator-metrics on namespace openshift-marketplace for network=default : 4.744363ms\\\\nI1122 04:48:21.325719 6483 services_controller.go:356] Processing sync for service openshift-kube-scheduler/scheduler for network=default\\\\nI1122 04:48:21.325674 6483 services_controller.go:452] Built service openshift-ingress-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325852 6483 services_controller.go:453] Built service openshift-ingress-operator/metrics template LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325922 6483 services_controller.go:454] Service openshift-ingress-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1122 04:48:21.325012 6483 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:44Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.113446 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.113488 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.113498 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.113513 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.113525 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:44Z","lastTransitionTime":"2025-11-22T04:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.115010 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:44Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.128011 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:44Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.139617 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:44Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.216649 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.216692 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.216702 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.216719 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.216730 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:44Z","lastTransitionTime":"2025-11-22T04:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.290127 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.290134 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:44 crc kubenswrapper[4818]: E1122 04:48:44.290337 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:44 crc kubenswrapper[4818]: E1122 04:48:44.290591 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.319181 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.319234 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.319275 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.319297 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.319313 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:44Z","lastTransitionTime":"2025-11-22T04:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.422621 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.422676 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.422692 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.422715 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.422733 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:44Z","lastTransitionTime":"2025-11-22T04:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.525155 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.525249 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.525313 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.525339 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.525356 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:44Z","lastTransitionTime":"2025-11-22T04:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.627768 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.627833 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.627852 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.627874 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.627892 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:44Z","lastTransitionTime":"2025-11-22T04:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.730535 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.730600 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.730617 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.730641 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.730658 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:44Z","lastTransitionTime":"2025-11-22T04:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.833312 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.833375 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.833394 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.833417 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.833433 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:44Z","lastTransitionTime":"2025-11-22T04:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.936232 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.936305 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.936321 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.936342 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:44 crc kubenswrapper[4818]: I1122 04:48:44.936359 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:44Z","lastTransitionTime":"2025-11-22T04:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.039932 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.039977 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.039990 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.040008 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.040021 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:45Z","lastTransitionTime":"2025-11-22T04:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.142979 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.143042 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.143060 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.143087 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.143109 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:45Z","lastTransitionTime":"2025-11-22T04:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.246198 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.246355 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.246377 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.246400 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.246418 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:45Z","lastTransitionTime":"2025-11-22T04:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.290098 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.290134 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:45 crc kubenswrapper[4818]: E1122 04:48:45.290312 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:45 crc kubenswrapper[4818]: E1122 04:48:45.290594 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.349526 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.349615 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.349634 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.349791 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.349855 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:45Z","lastTransitionTime":"2025-11-22T04:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.452935 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.452998 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.453015 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.453040 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.453057 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:45Z","lastTransitionTime":"2025-11-22T04:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.555893 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.555962 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.555983 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.556009 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.556026 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:45Z","lastTransitionTime":"2025-11-22T04:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.659330 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.659405 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.659427 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.659457 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.659481 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:45Z","lastTransitionTime":"2025-11-22T04:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.762530 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.762592 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.762613 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.762635 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.762650 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:45Z","lastTransitionTime":"2025-11-22T04:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.865940 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.866018 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.866045 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.866108 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.866145 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:45Z","lastTransitionTime":"2025-11-22T04:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.969387 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.969494 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.969527 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.969559 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:45 crc kubenswrapper[4818]: I1122 04:48:45.969585 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:45Z","lastTransitionTime":"2025-11-22T04:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.073092 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.073149 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.073166 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.073191 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.073210 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:46Z","lastTransitionTime":"2025-11-22T04:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.176537 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.176615 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.176633 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.176657 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.176675 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:46Z","lastTransitionTime":"2025-11-22T04:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.280360 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.280423 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.280440 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.280466 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.280485 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:46Z","lastTransitionTime":"2025-11-22T04:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.290836 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.290905 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:46 crc kubenswrapper[4818]: E1122 04:48:46.291059 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:46 crc kubenswrapper[4818]: E1122 04:48:46.291214 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.383939 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.384017 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.384036 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.384062 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.384081 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:46Z","lastTransitionTime":"2025-11-22T04:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.489986 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.490046 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.490062 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.490086 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.490104 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:46Z","lastTransitionTime":"2025-11-22T04:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.592810 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.592880 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.592900 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.592926 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.592944 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:46Z","lastTransitionTime":"2025-11-22T04:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.696687 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.696754 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.696768 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.696789 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.696806 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:46Z","lastTransitionTime":"2025-11-22T04:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.799905 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.799967 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.799995 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.800024 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.800047 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:46Z","lastTransitionTime":"2025-11-22T04:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.902548 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.902596 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.902608 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.902626 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:46 crc kubenswrapper[4818]: I1122 04:48:46.902639 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:46Z","lastTransitionTime":"2025-11-22T04:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.005868 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.005990 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.006015 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.006045 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.006066 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:47Z","lastTransitionTime":"2025-11-22T04:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.108600 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.108662 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.108683 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.108711 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.108732 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:47Z","lastTransitionTime":"2025-11-22T04:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.212706 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.212782 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.212801 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.212826 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.212845 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:47Z","lastTransitionTime":"2025-11-22T04:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.290922 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:47 crc kubenswrapper[4818]: E1122 04:48:47.291110 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.291154 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:47 crc kubenswrapper[4818]: E1122 04:48:47.291361 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.316213 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.316296 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.316312 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.316331 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.316344 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:47Z","lastTransitionTime":"2025-11-22T04:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.419144 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.419224 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.419241 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.419306 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.419333 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:47Z","lastTransitionTime":"2025-11-22T04:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.521684 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.522026 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.522166 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.522359 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.522501 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:47Z","lastTransitionTime":"2025-11-22T04:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.624930 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.625007 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.625031 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.625060 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.625081 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:47Z","lastTransitionTime":"2025-11-22T04:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.728826 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.728912 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.728951 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.728983 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.729005 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:47Z","lastTransitionTime":"2025-11-22T04:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.832489 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.832575 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.832604 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.832635 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.832658 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:47Z","lastTransitionTime":"2025-11-22T04:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.936343 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.936379 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.936390 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.936404 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:47 crc kubenswrapper[4818]: I1122 04:48:47.936416 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:47Z","lastTransitionTime":"2025-11-22T04:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.039109 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.039147 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.039165 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.039187 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.039204 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:48Z","lastTransitionTime":"2025-11-22T04:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.142452 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.142565 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.142585 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.142610 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.142626 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:48Z","lastTransitionTime":"2025-11-22T04:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.245638 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.245695 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.245713 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.245737 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.245757 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:48Z","lastTransitionTime":"2025-11-22T04:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.290692 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.290749 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:48 crc kubenswrapper[4818]: E1122 04:48:48.290902 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:48 crc kubenswrapper[4818]: E1122 04:48:48.291082 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.306302 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:48Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.324136 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:48Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.345790 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c569fd0932bade53c98da764d0d0e9dd07755d73dbba866c0be0209661aaa05a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:48Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.355436 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.355568 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.355597 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.355644 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.355668 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:48Z","lastTransitionTime":"2025-11-22T04:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.374344 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52d352c1148218337557f43c9772c5cc9866398ef20d8b5eb91eeff53c4cecbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"2025-11-22T04:47:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e6acc9f9-59f7-401f-8660-bd7d0fc9e096\\\\n2025-11-22T04:47:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e6acc9f9-59f7-401f-8660-bd7d0fc9e096 to /host/opt/cni/bin/\\\\n2025-11-22T04:47:57Z [verbose] multus-daemon started\\\\n2025-11-22T04:47:57Z [verbose] Readiness Indicator file check\\\\n2025-11-22T04:48:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:48Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.391091 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:48Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.409849 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"affc5026-facf-4e73-a63a-4bfe5ddf7268\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976c49c2e32cd03a00e6a0a91e325864c9a1eb48cc412cc03ed6cdd620da9125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e315f3e5e87e5ed47280e8efc909148c1b18af404778609c867f572a1c352c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe9de7b662c0ba4d848448a8322d3a464e58337895bf25390d4df018b695834a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:48Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.431062 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:48Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.450371 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:48Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.458539 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.458598 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.458615 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.458637 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.458653 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:48Z","lastTransitionTime":"2025-11-22T04:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.463593 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:48Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.477868 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:48Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.491897 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:48Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.505480 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:48Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.517645 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:48Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.538186 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:48Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.560954 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.561018 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.561036 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.561061 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.561079 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:48Z","lastTransitionTime":"2025-11-22T04:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.562104 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"e column _uuid == {de17f0de-cfb1-4534-bb42-c40f5e050c73}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1122 04:48:21.325074 6483 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}\\\\nI1122 04:48:21.325702 6483 services_controller.go:360] Finished syncing service marketplace-operator-metrics on namespace openshift-marketplace for network=default : 4.744363ms\\\\nI1122 04:48:21.325719 6483 services_controller.go:356] Processing sync for service openshift-kube-scheduler/scheduler for network=default\\\\nI1122 04:48:21.325674 6483 services_controller.go:452] Built service openshift-ingress-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325852 6483 services_controller.go:453] Built service openshift-ingress-operator/metrics template LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325922 6483 services_controller.go:454] Service openshift-ingress-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1122 04:48:21.325012 6483 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:48Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.584658 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:48Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.664785 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.664850 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.664868 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.664896 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.664914 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:48Z","lastTransitionTime":"2025-11-22T04:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.767628 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.767711 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.767734 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.767763 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.767781 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:48Z","lastTransitionTime":"2025-11-22T04:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.869690 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.869732 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.869745 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.869762 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.869776 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:48Z","lastTransitionTime":"2025-11-22T04:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.972549 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.972621 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.972638 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.972661 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:48 crc kubenswrapper[4818]: I1122 04:48:48.972678 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:48Z","lastTransitionTime":"2025-11-22T04:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.075839 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.075901 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.075920 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.075945 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.075963 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:49Z","lastTransitionTime":"2025-11-22T04:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.178613 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.178680 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.178700 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.178728 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.178750 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:49Z","lastTransitionTime":"2025-11-22T04:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.281021 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.281080 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.281101 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.281129 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.281167 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:49Z","lastTransitionTime":"2025-11-22T04:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.290532 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.290622 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:49 crc kubenswrapper[4818]: E1122 04:48:49.291061 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:49 crc kubenswrapper[4818]: E1122 04:48:49.291316 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.291712 4818 scope.go:117] "RemoveContainer" containerID="7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.384159 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.384214 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.384232 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.384281 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.384301 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:49Z","lastTransitionTime":"2025-11-22T04:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.487541 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.487601 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.487621 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.487645 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.487662 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:49Z","lastTransitionTime":"2025-11-22T04:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.589848 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.589897 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.589907 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.589924 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.589936 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:49Z","lastTransitionTime":"2025-11-22T04:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.695289 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.695338 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.695352 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.695368 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.695387 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:49Z","lastTransitionTime":"2025-11-22T04:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.812759 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.812820 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.812840 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.812868 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.812890 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:49Z","lastTransitionTime":"2025-11-22T04:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.916180 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.916297 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.916324 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.916353 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:49 crc kubenswrapper[4818]: I1122 04:48:49.916376 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:49Z","lastTransitionTime":"2025-11-22T04:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.019025 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.019108 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.019132 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.019162 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.019186 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:50Z","lastTransitionTime":"2025-11-22T04:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.122553 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.122629 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.122653 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.122679 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.122700 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:50Z","lastTransitionTime":"2025-11-22T04:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.226575 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.226635 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.226652 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.226676 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.226692 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:50Z","lastTransitionTime":"2025-11-22T04:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.291032 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.291151 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:50 crc kubenswrapper[4818]: E1122 04:48:50.291160 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:50 crc kubenswrapper[4818]: E1122 04:48:50.291364 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.329512 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.329573 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.329591 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.329613 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.329631 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:50Z","lastTransitionTime":"2025-11-22T04:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.432586 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.432646 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.432663 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.432686 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.432703 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:50Z","lastTransitionTime":"2025-11-22T04:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.535820 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.535886 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.535911 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.535940 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.535962 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:50Z","lastTransitionTime":"2025-11-22T04:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.638967 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.639045 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.639062 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.639081 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.639095 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:50Z","lastTransitionTime":"2025-11-22T04:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.742430 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.742850 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.742869 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.742895 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.742914 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:50Z","lastTransitionTime":"2025-11-22T04:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.847560 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.847632 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.847650 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.847674 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.847695 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:50Z","lastTransitionTime":"2025-11-22T04:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.861007 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p8s7r_f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/ovnkube-controller/2.log" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.865751 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerStarted","Data":"37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba"} Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.867333 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.890977 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:50Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.921388 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:50Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.949970 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.950034 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.950052 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.950075 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.950092 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:50Z","lastTransitionTime":"2025-11-22T04:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.962712 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"e column _uuid == {de17f0de-cfb1-4534-bb42-c40f5e050c73}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1122 04:48:21.325074 6483 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}\\\\nI1122 04:48:21.325702 6483 services_controller.go:360] Finished syncing service marketplace-operator-metrics on namespace openshift-marketplace for network=default : 4.744363ms\\\\nI1122 04:48:21.325719 6483 services_controller.go:356] Processing sync for service openshift-kube-scheduler/scheduler for network=default\\\\nI1122 04:48:21.325674 6483 services_controller.go:452] Built service openshift-ingress-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325852 6483 services_controller.go:453] Built service openshift-ingress-operator/metrics template LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325922 6483 services_controller.go:454] Service openshift-ingress-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1122 04:48:21.325012 6483 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:50Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:50 crc kubenswrapper[4818]: I1122 04:48:50.985809 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:50Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.001506 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:50Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.021176 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.035983 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52d352c1148218337557f43c9772c5cc9866398ef20d8b5eb91eeff53c4cecbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"2025-11-22T04:47:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e6acc9f9-59f7-401f-8660-bd7d0fc9e096\\\\n2025-11-22T04:47:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e6acc9f9-59f7-401f-8660-bd7d0fc9e096 to /host/opt/cni/bin/\\\\n2025-11-22T04:47:57Z [verbose] multus-daemon started\\\\n2025-11-22T04:47:57Z [verbose] Readiness Indicator file check\\\\n2025-11-22T04:48:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.052820 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.052859 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.052874 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.052889 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.052902 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:51Z","lastTransitionTime":"2025-11-22T04:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.059874 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c569fd0932bade53c98da764d0d0e9dd07755d73dbba866c0be0209661aaa05a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.076477 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.094167 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.113121 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.134138 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.152218 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.155138 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.155173 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.155186 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.155205 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.155218 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:51Z","lastTransitionTime":"2025-11-22T04:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.178383 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.191613 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.205750 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"affc5026-facf-4e73-a63a-4bfe5ddf7268\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976c49c2e32cd03a00e6a0a91e325864c9a1eb48cc412cc03ed6cdd620da9125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e315f3e5e87e5ed47280e8efc909148c1b18af404778609c867f572a1c352c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe9de7b662c0ba4d848448a8322d3a464e58337895bf25390d4df018b695834a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.257556 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.257598 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.257609 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.257626 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.257640 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:51Z","lastTransitionTime":"2025-11-22T04:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.290198 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:51 crc kubenswrapper[4818]: E1122 04:48:51.290374 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.290596 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:51 crc kubenswrapper[4818]: E1122 04:48:51.290809 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.360413 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.360474 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.360495 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.360519 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.360537 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:51Z","lastTransitionTime":"2025-11-22T04:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.466296 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.466699 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.466857 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.467004 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.467154 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:51Z","lastTransitionTime":"2025-11-22T04:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.570141 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.570214 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.570234 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.570287 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.570305 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:51Z","lastTransitionTime":"2025-11-22T04:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.673660 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.673725 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.673742 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.673774 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.673793 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:51Z","lastTransitionTime":"2025-11-22T04:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.776721 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.776789 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.776805 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.776828 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.776846 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:51Z","lastTransitionTime":"2025-11-22T04:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.873324 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p8s7r_f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/ovnkube-controller/3.log" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.874945 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p8s7r_f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/ovnkube-controller/2.log" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.879067 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.879109 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.879126 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.879148 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.879165 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:51Z","lastTransitionTime":"2025-11-22T04:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.880561 4818 generic.go:334] "Generic (PLEG): container finished" podID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerID="37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba" exitCode=1 Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.880644 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerDied","Data":"37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba"} Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.880756 4818 scope.go:117] "RemoveContainer" containerID="7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.881845 4818 scope.go:117] "RemoveContainer" containerID="37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba" Nov 22 04:48:51 crc kubenswrapper[4818]: E1122 04:48:51.882149 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.903477 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.922669 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.939844 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.954468 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.968529 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.983419 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.983471 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.983488 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.983512 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.983529 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:51Z","lastTransitionTime":"2025-11-22T04:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:51 crc kubenswrapper[4818]: I1122 04:48:51.989477 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"affc5026-facf-4e73-a63a-4bfe5ddf7268\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976c49c2e32cd03a00e6a0a91e325864c9a1eb48cc412cc03ed6cdd620da9125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e315f3e5e87e5ed47280e8efc909148c1b18af404778609c867f572a1c352c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe9de7b662c0ba4d848448a8322d3a464e58337895bf25390d4df018b695834a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.010446 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:52Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.030116 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:52Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.055870 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddd775f75f27ac42b15a145d6cdfa8b3b22ad4698393035b749396218a0f70b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:21Z\\\",\\\"message\\\":\\\"e column _uuid == {de17f0de-cfb1-4534-bb42-c40f5e050c73}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1122 04:48:21.325074 6483 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}\\\\nI1122 04:48:21.325702 6483 services_controller.go:360] Finished syncing service marketplace-operator-metrics on namespace openshift-marketplace for network=default : 4.744363ms\\\\nI1122 04:48:21.325719 6483 services_controller.go:356] Processing sync for service openshift-kube-scheduler/scheduler for network=default\\\\nI1122 04:48:21.325674 6483 services_controller.go:452] Built service openshift-ingress-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325852 6483 services_controller.go:453] Built service openshift-ingress-operator/metrics template LB for network=default: []services.LB{}\\\\nI1122 04:48:21.325922 6483 services_controller.go:454] Service openshift-ingress-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1122 04:48:21.325012 6483 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:51Z\\\",\\\"message\\\":\\\"y successful for *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc after 0 failed attempt(s)\\\\nI1122 04:48:51.400105 6881 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1122 04:48:51.400110 6881 default_network_controller.go:776] Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1122 04:48:51.400114 6881 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nF1122 04:48:51.399693 6881 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:52Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.080423 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:52Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.086199 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.086247 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.086326 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.086351 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.086368 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:52Z","lastTransitionTime":"2025-11-22T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.102165 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:52Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.123070 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:52Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.140398 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:52Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.158886 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:52Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.181964 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c569fd0932bade53c98da764d0d0e9dd07755d73dbba866c0be0209661aaa05a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:52Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.191509 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.191597 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.191616 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.191641 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.191657 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:52Z","lastTransitionTime":"2025-11-22T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.203917 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52d352c1148218337557f43c9772c5cc9866398ef20d8b5eb91eeff53c4cecbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"2025-11-22T04:47:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e6acc9f9-59f7-401f-8660-bd7d0fc9e096\\\\n2025-11-22T04:47:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e6acc9f9-59f7-401f-8660-bd7d0fc9e096 to /host/opt/cni/bin/\\\\n2025-11-22T04:47:57Z [verbose] multus-daemon started\\\\n2025-11-22T04:47:57Z [verbose] Readiness Indicator file check\\\\n2025-11-22T04:48:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:52Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.291043 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:52 crc kubenswrapper[4818]: E1122 04:48:52.291246 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.291367 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:52 crc kubenswrapper[4818]: E1122 04:48:52.291686 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.294549 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.294609 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.294627 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.294651 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.294670 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:52Z","lastTransitionTime":"2025-11-22T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.358896 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.358941 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.358956 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.358980 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.358997 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:52Z","lastTransitionTime":"2025-11-22T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:52 crc kubenswrapper[4818]: E1122 04:48:52.380362 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:52Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.385714 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.385770 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.385793 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.385825 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.385849 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:52Z","lastTransitionTime":"2025-11-22T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:52 crc kubenswrapper[4818]: E1122 04:48:52.406734 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:52Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.412109 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.412161 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.412179 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.412203 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.412223 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:52Z","lastTransitionTime":"2025-11-22T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:52 crc kubenswrapper[4818]: E1122 04:48:52.432426 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:52Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.438063 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.438124 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.438144 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.438171 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.438188 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:52Z","lastTransitionTime":"2025-11-22T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:52 crc kubenswrapper[4818]: E1122 04:48:52.459967 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:52Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.470503 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.470629 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.470990 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.471035 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.471060 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:52Z","lastTransitionTime":"2025-11-22T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:52 crc kubenswrapper[4818]: E1122 04:48:52.492647 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:52Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:52 crc kubenswrapper[4818]: E1122 04:48:52.492924 4818 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.495759 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.495835 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.495854 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.495879 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.495897 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:52Z","lastTransitionTime":"2025-11-22T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.599360 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.599422 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.599439 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.599466 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.599485 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:52Z","lastTransitionTime":"2025-11-22T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.702341 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.702439 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.702459 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.702484 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.702502 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:52Z","lastTransitionTime":"2025-11-22T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.805356 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.805413 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.805430 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.805455 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.805476 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:52Z","lastTransitionTime":"2025-11-22T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.887004 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p8s7r_f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/ovnkube-controller/3.log" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.892719 4818 scope.go:117] "RemoveContainer" containerID="37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba" Nov 22 04:48:52 crc kubenswrapper[4818]: E1122 04:48:52.892977 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.909021 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.909080 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.909101 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.909130 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.909152 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:52Z","lastTransitionTime":"2025-11-22T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.910299 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:52Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.926342 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:52Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.944793 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"affc5026-facf-4e73-a63a-4bfe5ddf7268\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976c49c2e32cd03a00e6a0a91e325864c9a1eb48cc412cc03ed6cdd620da9125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e315f3e5e87e5ed47280e8efc909148c1b18af404778609c867f572a1c352c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe9de7b662c0ba4d848448a8322d3a464e58337895bf25390d4df018b695834a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:52Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.967201 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:52Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:52 crc kubenswrapper[4818]: I1122 04:48:52.988998 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:52Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.006098 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:53Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.013196 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.013302 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.013324 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.013352 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.013374 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:53Z","lastTransitionTime":"2025-11-22T04:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.029421 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:53Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.046216 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:53Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.067549 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:53Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.087797 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:53Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.117939 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.118040 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.118058 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.118079 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.118095 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:53Z","lastTransitionTime":"2025-11-22T04:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.119270 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:51Z\\\",\\\"message\\\":\\\"y successful for *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc after 0 failed attempt(s)\\\\nI1122 04:48:51.400105 6881 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1122 04:48:51.400110 6881 default_network_controller.go:776] Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1122 04:48:51.400114 6881 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nF1122 04:48:51.399693 6881 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:53Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.139864 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:53Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.155216 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:53Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.171470 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:53Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.192504 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c569fd0932bade53c98da764d0d0e9dd07755d73dbba866c0be0209661aaa05a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:53Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.211469 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52d352c1148218337557f43c9772c5cc9866398ef20d8b5eb91eeff53c4cecbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"2025-11-22T04:47:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e6acc9f9-59f7-401f-8660-bd7d0fc9e096\\\\n2025-11-22T04:47:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e6acc9f9-59f7-401f-8660-bd7d0fc9e096 to /host/opt/cni/bin/\\\\n2025-11-22T04:47:57Z [verbose] multus-daemon started\\\\n2025-11-22T04:47:57Z [verbose] Readiness Indicator file check\\\\n2025-11-22T04:48:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:53Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.221810 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.221853 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.221869 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.221892 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.221910 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:53Z","lastTransitionTime":"2025-11-22T04:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.290591 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.290644 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:53 crc kubenswrapper[4818]: E1122 04:48:53.290804 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:53 crc kubenswrapper[4818]: E1122 04:48:53.290928 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.325312 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.325362 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.325377 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.325398 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.325414 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:53Z","lastTransitionTime":"2025-11-22T04:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.428175 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.428614 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.428836 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.429037 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.429195 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:53Z","lastTransitionTime":"2025-11-22T04:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.532644 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.533065 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.533209 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.533387 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.533527 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:53Z","lastTransitionTime":"2025-11-22T04:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.636328 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.636400 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.636424 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.636451 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.636469 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:53Z","lastTransitionTime":"2025-11-22T04:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.739984 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.740047 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.740070 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.740103 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.740125 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:53Z","lastTransitionTime":"2025-11-22T04:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.843083 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.843175 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.843200 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.843229 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.843284 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:53Z","lastTransitionTime":"2025-11-22T04:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.946911 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.946968 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.946985 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.947010 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:53 crc kubenswrapper[4818]: I1122 04:48:53.947027 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:53Z","lastTransitionTime":"2025-11-22T04:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.050522 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.050587 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.050608 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.050631 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.050649 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:54Z","lastTransitionTime":"2025-11-22T04:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.154748 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.154848 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.154875 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.154911 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.154947 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:54Z","lastTransitionTime":"2025-11-22T04:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.258614 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.258690 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.258702 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.258719 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.258730 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:54Z","lastTransitionTime":"2025-11-22T04:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.290952 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.291039 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:54 crc kubenswrapper[4818]: E1122 04:48:54.291654 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:54 crc kubenswrapper[4818]: E1122 04:48:54.291874 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.361798 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.361850 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.361867 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.361888 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.361904 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:54Z","lastTransitionTime":"2025-11-22T04:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.464812 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.465333 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.465483 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.465627 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.465765 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:54Z","lastTransitionTime":"2025-11-22T04:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.569223 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.569318 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.569337 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.569360 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.569378 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:54Z","lastTransitionTime":"2025-11-22T04:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.672456 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.672510 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.672527 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.672550 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.672567 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:54Z","lastTransitionTime":"2025-11-22T04:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.775138 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.775711 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.775926 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.776174 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.776461 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:54Z","lastTransitionTime":"2025-11-22T04:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.879481 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.879548 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.879572 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.879601 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.879622 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:54Z","lastTransitionTime":"2025-11-22T04:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.982714 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.982759 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.982771 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.982788 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:54 crc kubenswrapper[4818]: I1122 04:48:54.982800 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:54Z","lastTransitionTime":"2025-11-22T04:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.085746 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.085831 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.085845 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.085863 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.085874 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:55Z","lastTransitionTime":"2025-11-22T04:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.188319 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.188389 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.188407 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.188433 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.188452 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:55Z","lastTransitionTime":"2025-11-22T04:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.290483 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.290596 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:55 crc kubenswrapper[4818]: E1122 04:48:55.290670 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:55 crc kubenswrapper[4818]: E1122 04:48:55.290874 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.291601 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.291664 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.291683 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.291707 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.291724 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:55Z","lastTransitionTime":"2025-11-22T04:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.394578 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.394624 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.394633 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.394649 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.394661 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:55Z","lastTransitionTime":"2025-11-22T04:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.497170 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.497243 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.497280 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.497339 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.497377 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:55Z","lastTransitionTime":"2025-11-22T04:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.599678 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.599730 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.599743 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.599761 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.599777 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:55Z","lastTransitionTime":"2025-11-22T04:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.703288 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.703357 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.703374 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.703398 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.703416 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:55Z","lastTransitionTime":"2025-11-22T04:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.806708 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.806767 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.806787 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.806809 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.806826 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:55Z","lastTransitionTime":"2025-11-22T04:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.909740 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.909819 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.909839 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.909864 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:55 crc kubenswrapper[4818]: I1122 04:48:55.909884 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:55Z","lastTransitionTime":"2025-11-22T04:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.012748 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.012791 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.012805 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.012826 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.012839 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:56Z","lastTransitionTime":"2025-11-22T04:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.116764 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.116857 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.116883 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.116918 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.116944 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:56Z","lastTransitionTime":"2025-11-22T04:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.219905 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.220610 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.220856 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.221077 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.221240 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:56Z","lastTransitionTime":"2025-11-22T04:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.290122 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:56 crc kubenswrapper[4818]: E1122 04:48:56.290350 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.290129 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:56 crc kubenswrapper[4818]: E1122 04:48:56.290773 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.324585 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.324644 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.324661 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.324682 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.324701 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:56Z","lastTransitionTime":"2025-11-22T04:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.428926 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.429005 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.429026 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.429053 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.429081 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:56Z","lastTransitionTime":"2025-11-22T04:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.531825 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.531889 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.531908 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.531933 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.531953 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:56Z","lastTransitionTime":"2025-11-22T04:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.634908 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.634959 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.634975 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.634997 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.635014 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:56Z","lastTransitionTime":"2025-11-22T04:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.738187 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.738241 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.738285 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.738325 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.738363 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:56Z","lastTransitionTime":"2025-11-22T04:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.841735 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.841795 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.841813 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.841835 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.841854 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:56Z","lastTransitionTime":"2025-11-22T04:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.945338 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.945386 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.945399 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.945419 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:56 crc kubenswrapper[4818]: I1122 04:48:56.945431 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:56Z","lastTransitionTime":"2025-11-22T04:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.048616 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.048724 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.048745 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.048770 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.048787 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:57Z","lastTransitionTime":"2025-11-22T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.151650 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.151728 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.151751 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.151782 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.151828 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:57Z","lastTransitionTime":"2025-11-22T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.254329 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.254390 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.254408 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.254436 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.254457 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:57Z","lastTransitionTime":"2025-11-22T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.290467 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.290533 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:57 crc kubenswrapper[4818]: E1122 04:48:57.290601 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:57 crc kubenswrapper[4818]: E1122 04:48:57.290728 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.316764 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.316935 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:57 crc kubenswrapper[4818]: E1122 04:48:57.317068 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.317016135 +0000 UTC m=+153.891432742 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:48:57 crc kubenswrapper[4818]: E1122 04:48:57.317079 4818 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.317134 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:57 crc kubenswrapper[4818]: E1122 04:48:57.317189 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.317166609 +0000 UTC m=+153.891583316 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 04:48:57 crc kubenswrapper[4818]: E1122 04:48:57.317226 4818 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 04:48:57 crc kubenswrapper[4818]: E1122 04:48:57.317371 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.317348564 +0000 UTC m=+153.891765121 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.326331 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.357320 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.357358 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.357368 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.357384 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.357395 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:57Z","lastTransitionTime":"2025-11-22T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.418363 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.418454 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:57 crc kubenswrapper[4818]: E1122 04:48:57.418588 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 04:48:57 crc kubenswrapper[4818]: E1122 04:48:57.418621 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 04:48:57 crc kubenswrapper[4818]: E1122 04:48:57.418627 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 04:48:57 crc kubenswrapper[4818]: E1122 04:48:57.418638 4818 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:48:57 crc kubenswrapper[4818]: E1122 04:48:57.418653 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 04:48:57 crc kubenswrapper[4818]: E1122 04:48:57.418673 4818 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:48:57 crc kubenswrapper[4818]: E1122 04:48:57.418702 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.418684816 +0000 UTC m=+153.993101343 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:48:57 crc kubenswrapper[4818]: E1122 04:48:57.418738 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.418716867 +0000 UTC m=+153.993133434 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.459911 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.459963 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.459980 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.460005 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.460022 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:57Z","lastTransitionTime":"2025-11-22T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.561984 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.562033 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.562047 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.562066 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.562080 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:57Z","lastTransitionTime":"2025-11-22T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.665745 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.665804 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.665820 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.665844 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.665861 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:57Z","lastTransitionTime":"2025-11-22T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.769864 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.769954 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.769972 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.769995 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.770044 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:57Z","lastTransitionTime":"2025-11-22T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.872861 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.872954 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.872974 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.873000 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.873053 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:57Z","lastTransitionTime":"2025-11-22T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.976537 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.976632 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.976651 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.976677 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:57 crc kubenswrapper[4818]: I1122 04:48:57.976696 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:57Z","lastTransitionTime":"2025-11-22T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.079991 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.080080 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.080103 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.080131 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.080149 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:58Z","lastTransitionTime":"2025-11-22T04:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.183630 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.183706 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.183726 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.183751 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.183770 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:58Z","lastTransitionTime":"2025-11-22T04:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.286914 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.286976 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.286993 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.287019 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.287037 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:58Z","lastTransitionTime":"2025-11-22T04:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.291215 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.291324 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:48:58 crc kubenswrapper[4818]: E1122 04:48:58.291539 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:48:58 crc kubenswrapper[4818]: E1122 04:48:58.291685 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.314492 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.347574 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:51Z\\\",\\\"message\\\":\\\"y successful for *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc after 0 failed attempt(s)\\\\nI1122 04:48:51.400105 6881 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1122 04:48:51.400110 6881 default_network_controller.go:776] Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1122 04:48:51.400114 6881 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nF1122 04:48:51.399693 6881 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.371481 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.389059 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3aa0c6df-2d40-4c00-bab6-f23e2603e800\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9682e4f09980ee808df4b61b0bab8d8dd9d713a6211e5b3b6cf4ae6dffc84aa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab5d050214f11d3f38e205479fd500d3ef975f33f9026a5a49d9eb41d400cb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab5d050214f11d3f38e205479fd500d3ef975f33f9026a5a49d9eb41d400cb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.389735 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.389774 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.389790 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.389816 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.389838 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:58Z","lastTransitionTime":"2025-11-22T04:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.407990 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.436677 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.456096 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.479653 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52d352c1148218337557f43c9772c5cc9866398ef20d8b5eb91eeff53c4cecbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"2025-11-22T04:47:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e6acc9f9-59f7-401f-8660-bd7d0fc9e096\\\\n2025-11-22T04:47:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e6acc9f9-59f7-401f-8660-bd7d0fc9e096 to /host/opt/cni/bin/\\\\n2025-11-22T04:47:57Z [verbose] multus-daemon started\\\\n2025-11-22T04:47:57Z [verbose] Readiness Indicator file check\\\\n2025-11-22T04:48:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.492660 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.492735 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.492757 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.492786 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.492809 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:58Z","lastTransitionTime":"2025-11-22T04:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.503772 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c569fd0932bade53c98da764d0d0e9dd07755d73dbba866c0be0209661aaa05a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.522429 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.539736 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.558721 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.576330 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.593065 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.597018 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.597067 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.597086 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.597109 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.597127 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:58Z","lastTransitionTime":"2025-11-22T04:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.609547 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.628451 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"affc5026-facf-4e73-a63a-4bfe5ddf7268\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976c49c2e32cd03a00e6a0a91e325864c9a1eb48cc412cc03ed6cdd620da9125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e315f3e5e87e5ed47280e8efc909148c1b18af404778609c867f572a1c352c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe9de7b662c0ba4d848448a8322d3a464e58337895bf25390d4df018b695834a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.648413 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:58Z is after 2025-08-24T17:21:41Z" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.699305 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.699344 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.699354 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.699369 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.699380 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:58Z","lastTransitionTime":"2025-11-22T04:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.801960 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.802028 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.802055 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.802085 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.802108 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:58Z","lastTransitionTime":"2025-11-22T04:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.905070 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.905140 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.905160 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.905186 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:58 crc kubenswrapper[4818]: I1122 04:48:58.905204 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:58Z","lastTransitionTime":"2025-11-22T04:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.007700 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.007747 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.007764 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.007813 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.007831 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:59Z","lastTransitionTime":"2025-11-22T04:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.110642 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.110723 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.110754 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.110783 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.110804 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:59Z","lastTransitionTime":"2025-11-22T04:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.214234 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.214328 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.214345 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.214367 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.214384 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:59Z","lastTransitionTime":"2025-11-22T04:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.290515 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.290565 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:48:59 crc kubenswrapper[4818]: E1122 04:48:59.290989 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:48:59 crc kubenswrapper[4818]: E1122 04:48:59.291187 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.309532 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.317099 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.317184 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.317205 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.317274 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.317292 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:59Z","lastTransitionTime":"2025-11-22T04:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.420211 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.420296 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.420314 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.420335 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.420353 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:59Z","lastTransitionTime":"2025-11-22T04:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.523090 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.523146 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.523161 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.523186 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.523204 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:59Z","lastTransitionTime":"2025-11-22T04:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.626642 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.626706 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.626723 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.626747 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.626764 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:59Z","lastTransitionTime":"2025-11-22T04:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.729847 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.729914 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.729936 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.729965 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.729987 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:59Z","lastTransitionTime":"2025-11-22T04:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.835614 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.835681 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.835701 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.835725 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.835744 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:59Z","lastTransitionTime":"2025-11-22T04:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.938705 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.939130 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.939319 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.939486 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:48:59 crc kubenswrapper[4818]: I1122 04:48:59.939649 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:48:59Z","lastTransitionTime":"2025-11-22T04:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.043502 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.043579 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.043601 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.043629 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.043651 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:00Z","lastTransitionTime":"2025-11-22T04:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.147844 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.148334 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.148354 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.148382 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.148402 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:00Z","lastTransitionTime":"2025-11-22T04:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.252667 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.252734 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.252752 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.252781 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.252799 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:00Z","lastTransitionTime":"2025-11-22T04:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.291156 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.291156 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:00 crc kubenswrapper[4818]: E1122 04:49:00.291367 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:00 crc kubenswrapper[4818]: E1122 04:49:00.291501 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.356774 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.356839 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.356860 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.356888 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.356911 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:00Z","lastTransitionTime":"2025-11-22T04:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.460406 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.460704 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.460864 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.461079 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.461314 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:00Z","lastTransitionTime":"2025-11-22T04:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.563737 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.564091 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.564285 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.564440 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.564578 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:00Z","lastTransitionTime":"2025-11-22T04:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.667821 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.667910 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.667956 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.667978 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.667996 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:00Z","lastTransitionTime":"2025-11-22T04:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.770504 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.770575 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.770592 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.770618 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.770635 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:00Z","lastTransitionTime":"2025-11-22T04:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.873232 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.873317 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.873336 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.873359 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.873377 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:00Z","lastTransitionTime":"2025-11-22T04:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.976930 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.976993 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.977010 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.977034 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:00 crc kubenswrapper[4818]: I1122 04:49:00.977051 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:00Z","lastTransitionTime":"2025-11-22T04:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.079933 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.079991 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.080009 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.080032 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.080049 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:01Z","lastTransitionTime":"2025-11-22T04:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.182645 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.182689 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.182705 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.182726 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.182741 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:01Z","lastTransitionTime":"2025-11-22T04:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.285435 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.285499 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.285523 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.285552 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.285577 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:01Z","lastTransitionTime":"2025-11-22T04:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.290864 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:01 crc kubenswrapper[4818]: E1122 04:49:01.290996 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.290864 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:01 crc kubenswrapper[4818]: E1122 04:49:01.291099 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.389676 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.389738 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.389755 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.389777 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.389795 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:01Z","lastTransitionTime":"2025-11-22T04:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.493899 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.493971 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.493992 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.494016 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.494036 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:01Z","lastTransitionTime":"2025-11-22T04:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.596414 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.596487 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.596509 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.596541 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.596563 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:01Z","lastTransitionTime":"2025-11-22T04:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.699378 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.699443 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.699461 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.699511 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.699532 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:01Z","lastTransitionTime":"2025-11-22T04:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.803137 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.803206 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.803229 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.803289 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.803314 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:01Z","lastTransitionTime":"2025-11-22T04:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.905644 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.905691 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.905700 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.905712 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:01 crc kubenswrapper[4818]: I1122 04:49:01.905722 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:01Z","lastTransitionTime":"2025-11-22T04:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.008555 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.008610 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.008627 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.008649 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.008666 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:02Z","lastTransitionTime":"2025-11-22T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.112210 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.112336 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.112366 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.112397 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.112415 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:02Z","lastTransitionTime":"2025-11-22T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.214905 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.214977 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.215004 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.215033 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.215051 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:02Z","lastTransitionTime":"2025-11-22T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.290368 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.290434 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:02 crc kubenswrapper[4818]: E1122 04:49:02.290566 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:02 crc kubenswrapper[4818]: E1122 04:49:02.290726 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.321469 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.321529 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.321551 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.321579 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.321601 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:02Z","lastTransitionTime":"2025-11-22T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.424747 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.424781 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.424790 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.424803 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.424811 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:02Z","lastTransitionTime":"2025-11-22T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.527118 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.527180 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.527199 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.527223 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.527241 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:02Z","lastTransitionTime":"2025-11-22T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.629505 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.629544 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.629552 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.629566 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.629577 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:02Z","lastTransitionTime":"2025-11-22T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.705754 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.705828 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.705846 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.705875 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.705895 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:02Z","lastTransitionTime":"2025-11-22T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:02 crc kubenswrapper[4818]: E1122 04:49:02.726709 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:02Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.732288 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.732406 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.732426 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.732451 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.732468 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:02Z","lastTransitionTime":"2025-11-22T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:02 crc kubenswrapper[4818]: E1122 04:49:02.753143 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:02Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.757973 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.758035 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.758053 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.758075 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.758092 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:02Z","lastTransitionTime":"2025-11-22T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:02 crc kubenswrapper[4818]: E1122 04:49:02.777217 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:02Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.781836 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.781974 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.782024 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.782056 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.782074 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:02Z","lastTransitionTime":"2025-11-22T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:02 crc kubenswrapper[4818]: E1122 04:49:02.801294 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:02Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.806280 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.806330 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.806348 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.806372 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.806389 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:02Z","lastTransitionTime":"2025-11-22T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:02 crc kubenswrapper[4818]: E1122 04:49:02.825664 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:02Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:02 crc kubenswrapper[4818]: E1122 04:49:02.825884 4818 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.827904 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.827954 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.827972 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.827996 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.828013 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:02Z","lastTransitionTime":"2025-11-22T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.930843 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.930903 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.930922 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.930949 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:02 crc kubenswrapper[4818]: I1122 04:49:02.930969 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:02Z","lastTransitionTime":"2025-11-22T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.033879 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.033941 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.033960 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.033986 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.034009 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:03Z","lastTransitionTime":"2025-11-22T04:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.137315 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.137393 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.137421 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.137457 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.137482 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:03Z","lastTransitionTime":"2025-11-22T04:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.241013 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.241097 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.241123 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.241153 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.241174 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:03Z","lastTransitionTime":"2025-11-22T04:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.290483 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.290529 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:03 crc kubenswrapper[4818]: E1122 04:49:03.290671 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:03 crc kubenswrapper[4818]: E1122 04:49:03.290809 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.345086 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.345155 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.345172 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.345198 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.345216 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:03Z","lastTransitionTime":"2025-11-22T04:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.448467 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.448535 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.448553 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.448578 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.448599 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:03Z","lastTransitionTime":"2025-11-22T04:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.551997 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.552062 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.552084 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.552116 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.552136 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:03Z","lastTransitionTime":"2025-11-22T04:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.655887 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.655956 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.655983 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.656006 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.656022 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:03Z","lastTransitionTime":"2025-11-22T04:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.759218 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.759314 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.759338 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.759360 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.759378 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:03Z","lastTransitionTime":"2025-11-22T04:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.861964 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.862022 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.862039 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.862063 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.862080 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:03Z","lastTransitionTime":"2025-11-22T04:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.964759 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.964809 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.964826 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.964850 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:03 crc kubenswrapper[4818]: I1122 04:49:03.964866 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:03Z","lastTransitionTime":"2025-11-22T04:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.068242 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.068330 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.068346 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.068368 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.068385 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:04Z","lastTransitionTime":"2025-11-22T04:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.171850 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.171919 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.171949 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.172014 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.172036 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:04Z","lastTransitionTime":"2025-11-22T04:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.275260 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.275360 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.275383 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.275412 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.275433 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:04Z","lastTransitionTime":"2025-11-22T04:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.290930 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.291202 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:04 crc kubenswrapper[4818]: E1122 04:49:04.291397 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:04 crc kubenswrapper[4818]: E1122 04:49:04.291500 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.379246 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.379370 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.379410 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.379482 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.379508 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:04Z","lastTransitionTime":"2025-11-22T04:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.482203 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.482316 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.482342 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.482372 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.482395 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:04Z","lastTransitionTime":"2025-11-22T04:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.586452 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.586514 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.586539 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.586567 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.586589 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:04Z","lastTransitionTime":"2025-11-22T04:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.690136 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.690212 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.690231 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.690289 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.690307 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:04Z","lastTransitionTime":"2025-11-22T04:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.792852 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.792917 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.792939 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.792969 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.792991 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:04Z","lastTransitionTime":"2025-11-22T04:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.896395 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.896461 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.896486 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.896517 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:04 crc kubenswrapper[4818]: I1122 04:49:04.896540 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:04Z","lastTransitionTime":"2025-11-22T04:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.000909 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.001033 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.001062 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.001092 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.001113 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:05Z","lastTransitionTime":"2025-11-22T04:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.105347 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.105830 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.106054 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.106296 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.106477 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:05Z","lastTransitionTime":"2025-11-22T04:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.209201 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.209486 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.209652 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.209769 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.209855 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:05Z","lastTransitionTime":"2025-11-22T04:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.291060 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:05 crc kubenswrapper[4818]: E1122 04:49:05.291307 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.292316 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:05 crc kubenswrapper[4818]: E1122 04:49:05.292389 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.306643 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.312792 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.312846 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.312859 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.312876 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.312889 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:05Z","lastTransitionTime":"2025-11-22T04:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.415798 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.415857 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.415874 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.415898 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.415915 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:05Z","lastTransitionTime":"2025-11-22T04:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.518313 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.518352 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.518365 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.518381 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.518394 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:05Z","lastTransitionTime":"2025-11-22T04:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.621802 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.621864 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.621882 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.621907 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.621924 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:05Z","lastTransitionTime":"2025-11-22T04:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.725155 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.725219 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.725242 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.725326 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.725352 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:05Z","lastTransitionTime":"2025-11-22T04:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.827947 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.828039 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.828065 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.828096 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.828119 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:05Z","lastTransitionTime":"2025-11-22T04:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.931910 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.931972 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.931989 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.932014 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:05 crc kubenswrapper[4818]: I1122 04:49:05.932030 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:05Z","lastTransitionTime":"2025-11-22T04:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.035399 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.035449 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.035466 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.035488 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.035507 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:06Z","lastTransitionTime":"2025-11-22T04:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.138638 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.138964 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.138976 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.138991 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.139002 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:06Z","lastTransitionTime":"2025-11-22T04:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.241139 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.241201 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.241225 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.241289 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.241322 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:06Z","lastTransitionTime":"2025-11-22T04:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.290880 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:06 crc kubenswrapper[4818]: E1122 04:49:06.291052 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.291310 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:06 crc kubenswrapper[4818]: E1122 04:49:06.291497 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.345177 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.345320 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.345346 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.345376 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.345398 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:06Z","lastTransitionTime":"2025-11-22T04:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.449529 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.449597 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.449615 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.449640 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.449656 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:06Z","lastTransitionTime":"2025-11-22T04:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.552153 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.552209 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.552220 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.552237 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.552270 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:06Z","lastTransitionTime":"2025-11-22T04:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.655868 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.655914 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.655933 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.655957 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.655973 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:06Z","lastTransitionTime":"2025-11-22T04:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.759308 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.759372 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.759390 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.759414 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.759432 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:06Z","lastTransitionTime":"2025-11-22T04:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.862349 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.862420 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.862442 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.862465 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.862482 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:06Z","lastTransitionTime":"2025-11-22T04:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.964879 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.964942 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.964960 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.964984 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:06 crc kubenswrapper[4818]: I1122 04:49:06.965007 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:06Z","lastTransitionTime":"2025-11-22T04:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.068520 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.068591 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.068613 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.068640 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.068661 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:07Z","lastTransitionTime":"2025-11-22T04:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.171997 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.172073 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.172099 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.172133 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.172156 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:07Z","lastTransitionTime":"2025-11-22T04:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.275877 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.275932 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.275947 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.275969 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.275985 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:07Z","lastTransitionTime":"2025-11-22T04:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.290330 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.290350 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:07 crc kubenswrapper[4818]: E1122 04:49:07.290524 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:07 crc kubenswrapper[4818]: E1122 04:49:07.290648 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.379453 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.379519 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.379539 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.379564 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.379581 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:07Z","lastTransitionTime":"2025-11-22T04:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.482401 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.482450 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.482459 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.482476 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.482486 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:07Z","lastTransitionTime":"2025-11-22T04:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.585967 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.586040 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.586059 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.586086 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.586105 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:07Z","lastTransitionTime":"2025-11-22T04:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.689454 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.689518 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.689536 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.689562 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.689579 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:07Z","lastTransitionTime":"2025-11-22T04:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.792683 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.792747 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.792764 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.792790 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.792807 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:07Z","lastTransitionTime":"2025-11-22T04:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.895551 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.895646 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.895665 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.895689 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.895707 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:07Z","lastTransitionTime":"2025-11-22T04:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.998165 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.998226 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.998246 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.998299 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:07 crc kubenswrapper[4818]: I1122 04:49:07.998316 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:07Z","lastTransitionTime":"2025-11-22T04:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.100720 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.100786 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.100803 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.100829 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.100847 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:08Z","lastTransitionTime":"2025-11-22T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.203102 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.203156 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.203164 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.203177 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.203187 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:08Z","lastTransitionTime":"2025-11-22T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.290376 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:08 crc kubenswrapper[4818]: E1122 04:49:08.291063 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.291142 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.291611 4818 scope.go:117] "RemoveContainer" containerID="37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba" Nov 22 04:49:08 crc kubenswrapper[4818]: E1122 04:49:08.291695 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:08 crc kubenswrapper[4818]: E1122 04:49:08.291990 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.306305 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.306357 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.306387 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.306428 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.306446 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:08Z","lastTransitionTime":"2025-11-22T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.306685 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3aa0c6df-2d40-4c00-bab6-f23e2603e800\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9682e4f09980ee808df4b61b0bab8d8dd9d713a6211e5b3b6cf4ae6dffc84aa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab5d050214f11d3f38e205479fd500d3ef975f33f9026a5a49d9eb41d400cb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab5d050214f11d3f38e205479fd500d3ef975f33f9026a5a49d9eb41d400cb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.319691 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.339724 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.377202 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:51Z\\\",\\\"message\\\":\\\"y successful for *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc after 0 failed attempt(s)\\\\nI1122 04:48:51.400105 6881 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1122 04:48:51.400110 6881 default_network_controller.go:776] Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1122 04:48:51.400114 6881 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nF1122 04:48:51.399693 6881 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.397002 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.410027 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.410102 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.410127 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.410157 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.410178 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:08Z","lastTransitionTime":"2025-11-22T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.429675 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd76a674-9709-4eb6-84f4-77ec014ab7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://497c0b377c13ff186c85a1a98818a8aeb3adf0c705ae977e968bc8933b137934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d5fa59c436fa3b15782740e4c52b28f161dee88f7066780b0600abbac26871\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a760ca1879dd0698f20fe1a63f632243280be797c01ac221cd5b17dfd54f415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc48d49c0d054bc2d6af8010f53b799125ad4b3e67290d5c8da84b5a4a516e36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aadde793d0cd469559cedaa5d12ca8538604e9e400330380fda799d5abba734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc8845abe9f2530b18a2aba39b064bb0f7bdfb31dfa1331acb6499156db4ea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc8845abe9f2530b18a2aba39b064bb0f7bdfb31dfa1331acb6499156db4ea0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41515325c6456c08b79496fe17f63e7670d56ed4dc61ed276fb51b460ab5e87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41515325c6456c08b79496fe17f63e7670d56ed4dc61ed276fb51b460ab5e87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ddaa1fc41d78fa6bd011a178127d48fbec3859541b81be1c1fdde382e2777135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa1fc41d78fa6bd011a178127d48fbec3859541b81be1c1fdde382e2777135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.446830 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.464069 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.487818 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c569fd0932bade53c98da764d0d0e9dd07755d73dbba866c0be0209661aaa05a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.505574 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9776a326-749f-4d2e-837a-68000cacf7b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b13b8162d6f7654a176f320274869a5df27e278f935f0cab1068403f9e12f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://152036fde48e3fdcd33619dc1ca10977d41bb97c16bca4145926b4afe17267b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58d44809fc7aff83cf6c883f2db88130b98afdd8d29ea1a482560616853c207\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5690052f421eefbb317162b2ab5344182999498110d1bd671cc11116f4a8256b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.512960 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.513002 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.513019 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.513043 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.513059 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:08Z","lastTransitionTime":"2025-11-22T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.523141 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52d352c1148218337557f43c9772c5cc9866398ef20d8b5eb91eeff53c4cecbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"2025-11-22T04:47:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e6acc9f9-59f7-401f-8660-bd7d0fc9e096\\\\n2025-11-22T04:47:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e6acc9f9-59f7-401f-8660-bd7d0fc9e096 to /host/opt/cni/bin/\\\\n2025-11-22T04:47:57Z [verbose] multus-daemon started\\\\n2025-11-22T04:47:57Z [verbose] Readiness Indicator file check\\\\n2025-11-22T04:48:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.536486 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.554239 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"affc5026-facf-4e73-a63a-4bfe5ddf7268\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976c49c2e32cd03a00e6a0a91e325864c9a1eb48cc412cc03ed6cdd620da9125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e315f3e5e87e5ed47280e8efc909148c1b18af404778609c867f572a1c352c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe9de7b662c0ba4d848448a8322d3a464e58337895bf25390d4df018b695834a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.573471 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.593481 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.612634 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.616326 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.616390 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.616414 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.616442 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.616464 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:08Z","lastTransitionTime":"2025-11-22T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.635177 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.653036 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.668630 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:08Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.719928 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.719986 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.720003 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.720027 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.720044 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:08Z","lastTransitionTime":"2025-11-22T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.823108 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.823155 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.823168 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.823187 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.823202 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:08Z","lastTransitionTime":"2025-11-22T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.925838 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.925900 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.925917 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.925941 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:08 crc kubenswrapper[4818]: I1122 04:49:08.925959 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:08Z","lastTransitionTime":"2025-11-22T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.029143 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.029188 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.029200 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.029219 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.029231 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:09Z","lastTransitionTime":"2025-11-22T04:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.132912 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.132984 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.133003 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.133035 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.133058 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:09Z","lastTransitionTime":"2025-11-22T04:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.236693 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.236752 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.236771 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.236808 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.236833 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:09Z","lastTransitionTime":"2025-11-22T04:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.290072 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.290089 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:09 crc kubenswrapper[4818]: E1122 04:49:09.290453 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:09 crc kubenswrapper[4818]: E1122 04:49:09.290563 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.340772 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.340805 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.340816 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.340830 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.340841 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:09Z","lastTransitionTime":"2025-11-22T04:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.444652 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.445443 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.445482 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.445509 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.445530 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:09Z","lastTransitionTime":"2025-11-22T04:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.548478 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.548517 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.548530 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.548549 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.548561 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:09Z","lastTransitionTime":"2025-11-22T04:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.650987 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.651023 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.651034 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.651050 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.651061 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:09Z","lastTransitionTime":"2025-11-22T04:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.753764 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.753808 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.753822 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.753842 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.753857 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:09Z","lastTransitionTime":"2025-11-22T04:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.856937 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.857007 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.857041 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.857074 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.857098 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:09Z","lastTransitionTime":"2025-11-22T04:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.960045 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.960111 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.960129 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.960157 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:09 crc kubenswrapper[4818]: I1122 04:49:09.960176 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:09Z","lastTransitionTime":"2025-11-22T04:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.063849 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.063988 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.064012 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.064037 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.064056 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:10Z","lastTransitionTime":"2025-11-22T04:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.167114 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.167191 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.167227 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.167289 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.167312 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:10Z","lastTransitionTime":"2025-11-22T04:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.269885 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.269938 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.269949 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.269966 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.269977 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:10Z","lastTransitionTime":"2025-11-22T04:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.290606 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.290680 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:10 crc kubenswrapper[4818]: E1122 04:49:10.290765 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:10 crc kubenswrapper[4818]: E1122 04:49:10.290879 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.371399 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs\") pod \"network-metrics-daemon-djd77\" (UID: \"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\") " pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:10 crc kubenswrapper[4818]: E1122 04:49:10.371619 4818 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 04:49:10 crc kubenswrapper[4818]: E1122 04:49:10.371752 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs podName:429a37e7-cb52-4e76-89ee-16f2b9e6fbfe nodeName:}" failed. No retries permitted until 2025-11-22 04:50:14.371722458 +0000 UTC m=+166.946139015 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs") pod "network-metrics-daemon-djd77" (UID: "429a37e7-cb52-4e76-89ee-16f2b9e6fbfe") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.372635 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.372693 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.372710 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.372734 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.372751 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:10Z","lastTransitionTime":"2025-11-22T04:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.475279 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.475332 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.475343 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.475361 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.475373 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:10Z","lastTransitionTime":"2025-11-22T04:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.578203 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.578318 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.578340 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.578365 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.578383 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:10Z","lastTransitionTime":"2025-11-22T04:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.681205 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.681297 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.681316 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.681339 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.681356 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:10Z","lastTransitionTime":"2025-11-22T04:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.784713 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.784783 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.784804 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.784826 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.784845 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:10Z","lastTransitionTime":"2025-11-22T04:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.890792 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.890853 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.890872 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.890901 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.890928 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:10Z","lastTransitionTime":"2025-11-22T04:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.994601 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.994716 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.994771 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.994801 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:10 crc kubenswrapper[4818]: I1122 04:49:10.994819 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:10Z","lastTransitionTime":"2025-11-22T04:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.097959 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.098024 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.098043 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.098068 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.098086 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:11Z","lastTransitionTime":"2025-11-22T04:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.201325 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.201389 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.201406 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.201429 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.201446 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:11Z","lastTransitionTime":"2025-11-22T04:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.290711 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.290776 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:11 crc kubenswrapper[4818]: E1122 04:49:11.290922 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:11 crc kubenswrapper[4818]: E1122 04:49:11.291049 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.304336 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.304427 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.304437 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.304458 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.304474 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:11Z","lastTransitionTime":"2025-11-22T04:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.408578 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.408643 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.408660 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.408685 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.408702 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:11Z","lastTransitionTime":"2025-11-22T04:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.511435 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.511495 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.511512 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.511537 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.511554 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:11Z","lastTransitionTime":"2025-11-22T04:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.614487 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.614569 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.614592 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.614620 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.614638 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:11Z","lastTransitionTime":"2025-11-22T04:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.717988 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.718063 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.718099 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.718129 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.718150 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:11Z","lastTransitionTime":"2025-11-22T04:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.821046 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.821123 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.821142 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.821165 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.821181 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:11Z","lastTransitionTime":"2025-11-22T04:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.924571 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.924657 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.924688 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.924717 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:11 crc kubenswrapper[4818]: I1122 04:49:11.924737 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:11Z","lastTransitionTime":"2025-11-22T04:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.027389 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.027475 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.027503 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.027531 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.027554 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:12Z","lastTransitionTime":"2025-11-22T04:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.130161 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.130224 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.130240 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.130306 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.130326 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:12Z","lastTransitionTime":"2025-11-22T04:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.233519 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.233579 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.233598 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.233620 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.233637 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:12Z","lastTransitionTime":"2025-11-22T04:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.290621 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:12 crc kubenswrapper[4818]: E1122 04:49:12.290838 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.290906 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:12 crc kubenswrapper[4818]: E1122 04:49:12.291112 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.336497 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.336557 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.336574 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.336596 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.336613 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:12Z","lastTransitionTime":"2025-11-22T04:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.440027 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.440096 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.440114 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.440137 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.440156 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:12Z","lastTransitionTime":"2025-11-22T04:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.544185 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.544246 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.544305 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.544334 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.544361 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:12Z","lastTransitionTime":"2025-11-22T04:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.647129 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.647193 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.647210 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.647235 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.647285 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:12Z","lastTransitionTime":"2025-11-22T04:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.751600 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.751657 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.751676 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.751702 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.751720 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:12Z","lastTransitionTime":"2025-11-22T04:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.855472 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.855563 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.855592 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.855625 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.855649 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:12Z","lastTransitionTime":"2025-11-22T04:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.958756 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.958828 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.958845 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.958872 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:12 crc kubenswrapper[4818]: I1122 04:49:12.958947 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:12Z","lastTransitionTime":"2025-11-22T04:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.061629 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.061707 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.061724 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.061754 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.061772 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:13Z","lastTransitionTime":"2025-11-22T04:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.131294 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.131351 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.131367 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.131389 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.131413 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:13Z","lastTransitionTime":"2025-11-22T04:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:13 crc kubenswrapper[4818]: E1122 04:49:13.151889 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:13Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.157445 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.157494 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.157511 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.157534 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.157551 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:13Z","lastTransitionTime":"2025-11-22T04:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:13 crc kubenswrapper[4818]: E1122 04:49:13.179664 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:13Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.185082 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.185137 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.185156 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.185179 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.185195 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:13Z","lastTransitionTime":"2025-11-22T04:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:13 crc kubenswrapper[4818]: E1122 04:49:13.205740 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:13Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.210652 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.210693 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.210708 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.210731 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.210747 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:13Z","lastTransitionTime":"2025-11-22T04:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:13 crc kubenswrapper[4818]: E1122 04:49:13.230904 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:13Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.235612 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.235665 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.235681 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.235705 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.235722 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:13Z","lastTransitionTime":"2025-11-22T04:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:13 crc kubenswrapper[4818]: E1122 04:49:13.257049 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:13Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:13 crc kubenswrapper[4818]: E1122 04:49:13.257453 4818 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.259856 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.259973 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.260001 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.260066 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.260087 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:13Z","lastTransitionTime":"2025-11-22T04:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.291164 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.291535 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:13 crc kubenswrapper[4818]: E1122 04:49:13.291674 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:13 crc kubenswrapper[4818]: E1122 04:49:13.291956 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.362899 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.362969 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.362987 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.363011 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.363030 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:13Z","lastTransitionTime":"2025-11-22T04:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.465291 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.465364 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.465388 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.465417 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.465440 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:13Z","lastTransitionTime":"2025-11-22T04:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.567983 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.568075 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.568093 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.568118 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.568137 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:13Z","lastTransitionTime":"2025-11-22T04:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.671644 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.671706 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.671725 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.671748 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.671766 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:13Z","lastTransitionTime":"2025-11-22T04:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.774934 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.775090 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.775119 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.775147 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.775168 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:13Z","lastTransitionTime":"2025-11-22T04:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.877885 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.877948 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.877970 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.878000 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.878020 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:13Z","lastTransitionTime":"2025-11-22T04:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.980622 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.980690 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.980711 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.980735 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:13 crc kubenswrapper[4818]: I1122 04:49:13.980753 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:13Z","lastTransitionTime":"2025-11-22T04:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.083890 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.083988 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.084007 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.084067 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.084086 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:14Z","lastTransitionTime":"2025-11-22T04:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.187771 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.187850 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.187868 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.187890 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.187939 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:14Z","lastTransitionTime":"2025-11-22T04:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.290615 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:14 crc kubenswrapper[4818]: E1122 04:49:14.290824 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.290615 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:14 crc kubenswrapper[4818]: E1122 04:49:14.291519 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.291925 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.292006 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.292024 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.292048 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.292066 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:14Z","lastTransitionTime":"2025-11-22T04:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.395176 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.395293 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.395320 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.395354 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.395373 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:14Z","lastTransitionTime":"2025-11-22T04:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.498440 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.498495 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.498512 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.498536 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.498555 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:14Z","lastTransitionTime":"2025-11-22T04:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.602386 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.603342 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.603535 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.603670 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.603845 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:14Z","lastTransitionTime":"2025-11-22T04:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.707003 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.707092 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.707119 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.707148 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.707169 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:14Z","lastTransitionTime":"2025-11-22T04:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.810014 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.810384 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.810566 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.810712 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.810836 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:14Z","lastTransitionTime":"2025-11-22T04:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.913712 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.913767 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.913786 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.913811 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:14 crc kubenswrapper[4818]: I1122 04:49:14.913829 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:14Z","lastTransitionTime":"2025-11-22T04:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.016927 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.016995 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.017020 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.017051 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.017071 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:15Z","lastTransitionTime":"2025-11-22T04:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.120144 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.120215 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.120232 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.120288 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.120306 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:15Z","lastTransitionTime":"2025-11-22T04:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.223350 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.223397 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.223407 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.223424 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.223435 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:15Z","lastTransitionTime":"2025-11-22T04:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.290694 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:15 crc kubenswrapper[4818]: E1122 04:49:15.290841 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.290694 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:15 crc kubenswrapper[4818]: E1122 04:49:15.291052 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.326054 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.326353 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.326507 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.326612 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.326699 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:15Z","lastTransitionTime":"2025-11-22T04:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.430061 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.430110 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.430122 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.430139 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.430149 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:15Z","lastTransitionTime":"2025-11-22T04:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.533367 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.533429 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.533450 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.533478 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.533501 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:15Z","lastTransitionTime":"2025-11-22T04:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.636864 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.636956 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.636981 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.637010 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.637031 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:15Z","lastTransitionTime":"2025-11-22T04:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.740189 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.740241 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.740288 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.740311 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.740330 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:15Z","lastTransitionTime":"2025-11-22T04:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.844114 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.844178 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.844199 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.844228 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.844289 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:15Z","lastTransitionTime":"2025-11-22T04:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.948574 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.948731 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.948753 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.948778 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:15 crc kubenswrapper[4818]: I1122 04:49:15.948797 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:15Z","lastTransitionTime":"2025-11-22T04:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.051739 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.051788 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.051804 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.051827 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.051848 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:16Z","lastTransitionTime":"2025-11-22T04:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.155339 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.155452 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.155479 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.155510 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.155533 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:16Z","lastTransitionTime":"2025-11-22T04:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.259642 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.259691 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.259707 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.259727 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.259740 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:16Z","lastTransitionTime":"2025-11-22T04:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.291065 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.291101 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:16 crc kubenswrapper[4818]: E1122 04:49:16.291349 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:16 crc kubenswrapper[4818]: E1122 04:49:16.291480 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.367799 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.368079 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.368150 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.368213 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.368349 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:16Z","lastTransitionTime":"2025-11-22T04:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.471461 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.471525 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.471542 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.471565 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.471588 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:16Z","lastTransitionTime":"2025-11-22T04:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.575126 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.575223 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.575247 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.575318 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.575342 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:16Z","lastTransitionTime":"2025-11-22T04:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.686319 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.686400 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.686468 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.686492 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.686509 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:16Z","lastTransitionTime":"2025-11-22T04:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.790137 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.790202 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.790219 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.790245 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.790288 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:16Z","lastTransitionTime":"2025-11-22T04:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.893896 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.894029 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.894055 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.894084 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.894144 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:16Z","lastTransitionTime":"2025-11-22T04:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.996866 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.996937 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.996960 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.996990 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:16 crc kubenswrapper[4818]: I1122 04:49:16.997013 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:16Z","lastTransitionTime":"2025-11-22T04:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.101096 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.101179 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.101203 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.101234 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.101294 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:17Z","lastTransitionTime":"2025-11-22T04:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.203771 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.203838 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.203856 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.203880 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.203898 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:17Z","lastTransitionTime":"2025-11-22T04:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.290520 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.290571 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:17 crc kubenswrapper[4818]: E1122 04:49:17.291018 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:17 crc kubenswrapper[4818]: E1122 04:49:17.291040 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.307095 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.307202 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.307224 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.307315 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.307337 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:17Z","lastTransitionTime":"2025-11-22T04:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.410133 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.410217 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.410290 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.410327 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.410349 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:17Z","lastTransitionTime":"2025-11-22T04:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.513898 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.513963 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.513999 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.514033 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.514059 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:17Z","lastTransitionTime":"2025-11-22T04:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.617809 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.617852 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.617864 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.617880 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.617891 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:17Z","lastTransitionTime":"2025-11-22T04:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.721929 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.721991 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.722008 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.722034 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.722052 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:17Z","lastTransitionTime":"2025-11-22T04:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.825824 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.825890 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.825925 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.825950 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.825969 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:17Z","lastTransitionTime":"2025-11-22T04:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.928892 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.928967 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.928992 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.929022 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:17 crc kubenswrapper[4818]: I1122 04:49:17.929046 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:17Z","lastTransitionTime":"2025-11-22T04:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.032123 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.032182 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.032200 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.032222 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.032242 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:18Z","lastTransitionTime":"2025-11-22T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.135774 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.135832 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.135847 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.135865 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.135877 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:18Z","lastTransitionTime":"2025-11-22T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.238893 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.238938 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.238954 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.238976 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.238990 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:18Z","lastTransitionTime":"2025-11-22T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.290998 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.291039 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:18 crc kubenswrapper[4818]: E1122 04:49:18.291210 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:18 crc kubenswrapper[4818]: E1122 04:49:18.291370 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.324464 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:51Z\\\",\\\"message\\\":\\\"y successful for *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc after 0 failed attempt(s)\\\\nI1122 04:48:51.400105 6881 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1122 04:48:51.400110 6881 default_network_controller.go:776] Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1122 04:48:51.400114 6881 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nF1122 04:48:51.399693 6881 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:48:51Z is after 2025-08-24\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-97n69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p8s7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.343767 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.343840 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.343863 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.343893 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.343913 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:18Z","lastTransitionTime":"2025-11-22T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.350167 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f281330-68fd-4866-a787-83440a58aa4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://848a04def65dfbbef8aa0b8d7e7872024c65816fd930bd0cad8fb214ed584f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef8bd928ef35b683a12c946e4d75075306a1662fe65ee18cfa3401bc11596a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bd197510ea3dc8881aae41de1c5b15b448f5bd6b64699c753bebe665249dbb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ad4435eff90aaf2e68f6d46518be7bfd7ab5d126279cc6273e4931dc72cd08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0358e913df2f7564ffbaba7d624fc5a0ffcdb0d2ca99603b48494691a19c2a8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://436b902a8dbb8f6277bd970df9013827d8fce062b909987131fe1456bcfdd3e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c635ab5a7a53c1b07032ea515261f86984eff4dc48182f86007d50df7f476cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv2fh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9r6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.366823 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3aa0c6df-2d40-4c00-bab6-f23e2603e800\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9682e4f09980ee808df4b61b0bab8d8dd9d713a6211e5b3b6cf4ae6dffc84aa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab5d050214f11d3f38e205479fd500d3ef975f33f9026a5a49d9eb41d400cb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab5d050214f11d3f38e205479fd500d3ef975f33f9026a5a49d9eb41d400cb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.388441 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.409131 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b615bfc03380785e7f5c4215d28c92f6ceaa24cc60dd98adb8fd7bf586492082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee59c84df705e0b85cd6dc7c10a3718396d1a9993065a30967841446855297f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.432657 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd76a674-9709-4eb6-84f4-77ec014ab7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://497c0b377c13ff186c85a1a98818a8aeb3adf0c705ae977e968bc8933b137934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d5fa59c436fa3b15782740e4c52b28f161dee88f7066780b0600abbac26871\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a760ca1879dd0698f20fe1a63f632243280be797c01ac221cd5b17dfd54f415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc48d49c0d054bc2d6af8010f53b799125ad4b3e67290d5c8da84b5a4a516e36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aadde793d0cd469559cedaa5d12ca8538604e9e400330380fda799d5abba734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc8845abe9f2530b18a2aba39b064bb0f7bdfb31dfa1331acb6499156db4ea0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc8845abe9f2530b18a2aba39b064bb0f7bdfb31dfa1331acb6499156db4ea0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41515325c6456c08b79496fe17f63e7670d56ed4dc61ed276fb51b460ab5e87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41515325c6456c08b79496fe17f63e7670d56ed4dc61ed276fb51b460ab5e87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ddaa1fc41d78fa6bd011a178127d48fbec3859541b81be1c1fdde382e2777135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddaa1fc41d78fa6bd011a178127d48fbec3859541b81be1c1fdde382e2777135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.443227 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ct4xj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"503d7b0b-1218-4c3b-b746-baea82235e2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90acaa1f4ac5403764363e3e9fea9b0a91362792e787f3de3a6c6e81c3e46b91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89htt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ct4xj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.445854 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.445909 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.445921 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.445941 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.445955 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:18Z","lastTransitionTime":"2025-11-22T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.453098 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fda14-7029-4dfe-adf4-7b3301d926b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a2403c53fc8a2744709dd0a83e8dbf91bc3983a879785234865cdb9985bca3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9870147bc0268ef9941cc978e9b07cba5f99f7d592383ea892b902bc78284049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kl2rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x52mk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.469682 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a9c8bbf-972c-4c7c-8f28-e347d9936d12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4526377420a470f29cbc3fda654c3b29875e2448825aee2164945ffe40c9e7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://378af87501021f2531b406087501ff1b35bacf1e299692e0c8ff00af24ca26b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbe25f24d3b4ccf6ba4e45f7998756b69960e9bdfa9a527f8cce0da895095bb4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c569fd0932bade53c98da764d0d0e9dd07755d73dbba866c0be0209661aaa05a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1740a1ffaa58efe2b9bf6733bcc6aa5ef2ad89aa967d6900d3d4f8c05061abe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1122 04:47:45.180209 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 04:47:45.181335 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3232551165/tls.crt::/tmp/serving-cert-3232551165/tls.key\\\\\\\"\\\\nI1122 04:47:50.947977 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 04:47:50.953872 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 04:47:50.953917 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 04:47:50.953962 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 04:47:50.953977 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 04:47:50.982484 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 04:47:50.982528 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 04:47:50.982539 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1122 04:47:50.982538 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 04:47:50.982552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 04:47:50.982586 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 04:47:50.982593 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 04:47:50.982599 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 04:47:50.986623 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d284a13091ed4269c7e89171a803470300a22dc9e45c70798e8bd31422ded2e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca413aa6d0b00226c2af9ad6a69a47817349a37099b6102009f777de43be6f3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.481934 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9776a326-749f-4d2e-837a-68000cacf7b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b13b8162d6f7654a176f320274869a5df27e278f935f0cab1068403f9e12f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://152036fde48e3fdcd33619dc1ca10977d41bb97c16bca4145926b4afe17267b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58d44809fc7aff83cf6c883f2db88130b98afdd8d29ea1a482560616853c207\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5690052f421eefbb317162b2ab5344182999498110d1bd671cc11116f4a8256b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.499400 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x865d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d96879f8-2766-4f5e-bc3e-bbf8e5394a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52d352c1148218337557f43c9772c5cc9866398ef20d8b5eb91eeff53c4cecbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T04:48:42Z\\\",\\\"message\\\":\\\"2025-11-22T04:47:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e6acc9f9-59f7-401f-8660-bd7d0fc9e096\\\\n2025-11-22T04:47:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e6acc9f9-59f7-401f-8660-bd7d0fc9e096 to /host/opt/cni/bin/\\\\n2025-11-22T04:47:57Z [verbose] multus-daemon started\\\\n2025-11-22T04:47:57Z [verbose] Readiness Indicator file check\\\\n2025-11-22T04:48:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:48:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwd55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x865d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.513666 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac32c4ca95d487177cd1a508c3629dcb2934bbba3bab760b62946ecb0b5afed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.528559 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.542648 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0971c9ec-29ce-44eb-9875-ad4664143305\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e41b0e907fffec90d0cb27a275eb45544415739542b51fc0c206d6a9a982aad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b95fx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rjvw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.548686 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.548712 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.548720 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.548735 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.548744 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:18Z","lastTransitionTime":"2025-11-22T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.553835 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n4wst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84fcc74a-fc3b-4885-9c82-6f3956d50c92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa177315303cc9702018fe5c6c3976a69ab1966344ee0ec2c9df2d11c5b4dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ff46s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n4wst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.566726 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-djd77" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:48:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-djd77\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.581633 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"affc5026-facf-4e73-a63a-4bfe5ddf7268\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976c49c2e32cd03a00e6a0a91e325864c9a1eb48cc412cc03ed6cdd620da9125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e315f3e5e87e5ed47280e8efc909148c1b18af404778609c867f572a1c352c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe9de7b662c0ba4d848448a8322d3a464e58337895bf25390d4df018b695834a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f845943018499c6457f3385a158de6e9293f3dc29b499378583ce472db368f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T04:47:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T04:47:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.594960 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.614733 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T04:47:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a72ea403fe1a9fde108a31f147fc61ccc7659137334fe144c729881c4b6c9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T04:47:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:18Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.651588 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.651636 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.651654 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.651692 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.651712 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:18Z","lastTransitionTime":"2025-11-22T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.754300 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.754339 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.754352 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.754368 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.754379 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:18Z","lastTransitionTime":"2025-11-22T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.857502 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.857564 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.857582 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.857703 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.857724 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:18Z","lastTransitionTime":"2025-11-22T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.959837 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.959904 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.959926 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.959955 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:18 crc kubenswrapper[4818]: I1122 04:49:18.959976 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:18Z","lastTransitionTime":"2025-11-22T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.062177 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.062228 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.062238 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.062264 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.062273 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:19Z","lastTransitionTime":"2025-11-22T04:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.165135 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.165289 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.165309 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.165333 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.165349 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:19Z","lastTransitionTime":"2025-11-22T04:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.268619 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.268690 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.268708 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.268736 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.268755 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:19Z","lastTransitionTime":"2025-11-22T04:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.290464 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.290555 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:19 crc kubenswrapper[4818]: E1122 04:49:19.290625 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:19 crc kubenswrapper[4818]: E1122 04:49:19.290713 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.372708 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.372797 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.372823 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.372855 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.372880 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:19Z","lastTransitionTime":"2025-11-22T04:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.475551 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.475606 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.475626 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.475651 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.475668 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:19Z","lastTransitionTime":"2025-11-22T04:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.579310 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.579411 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.579428 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.579455 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.579476 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:19Z","lastTransitionTime":"2025-11-22T04:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.682468 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.682596 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.682615 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.682637 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.682654 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:19Z","lastTransitionTime":"2025-11-22T04:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.785841 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.785891 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.785907 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.785934 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.785950 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:19Z","lastTransitionTime":"2025-11-22T04:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.888954 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.889014 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.889031 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.889056 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.889074 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:19Z","lastTransitionTime":"2025-11-22T04:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.992348 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.992422 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.992445 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.992473 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:19 crc kubenswrapper[4818]: I1122 04:49:19.992497 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:19Z","lastTransitionTime":"2025-11-22T04:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.094938 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.094987 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.094997 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.095012 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.095022 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:20Z","lastTransitionTime":"2025-11-22T04:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.197278 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.197313 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.197321 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.197334 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.197342 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:20Z","lastTransitionTime":"2025-11-22T04:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.290656 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:20 crc kubenswrapper[4818]: E1122 04:49:20.290790 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.290662 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:20 crc kubenswrapper[4818]: E1122 04:49:20.290939 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.299682 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.299761 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.299773 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.299814 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.299828 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:20Z","lastTransitionTime":"2025-11-22T04:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.402248 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.402369 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.402389 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.402416 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.402432 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:20Z","lastTransitionTime":"2025-11-22T04:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.506371 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.506441 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.506468 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.506497 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.506518 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:20Z","lastTransitionTime":"2025-11-22T04:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.609513 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.609556 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.609568 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.609586 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.609603 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:20Z","lastTransitionTime":"2025-11-22T04:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.713166 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.713271 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.713284 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.713302 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.713315 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:20Z","lastTransitionTime":"2025-11-22T04:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.816767 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.817419 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.817451 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.817482 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.817503 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:20Z","lastTransitionTime":"2025-11-22T04:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.920877 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.920939 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.920957 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.920979 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:20 crc kubenswrapper[4818]: I1122 04:49:20.920996 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:20Z","lastTransitionTime":"2025-11-22T04:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.023773 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.024008 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.024157 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.024347 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.024490 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:21Z","lastTransitionTime":"2025-11-22T04:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.165159 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.165210 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.165233 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.165291 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.165316 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:21Z","lastTransitionTime":"2025-11-22T04:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.268473 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.268841 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.268895 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.268927 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.268945 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:21Z","lastTransitionTime":"2025-11-22T04:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.290929 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.290936 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:21 crc kubenswrapper[4818]: E1122 04:49:21.291115 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:21 crc kubenswrapper[4818]: E1122 04:49:21.291303 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.372462 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.372679 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.372810 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.372949 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.373086 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:21Z","lastTransitionTime":"2025-11-22T04:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.476021 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.476078 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.476096 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.476119 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.476139 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:21Z","lastTransitionTime":"2025-11-22T04:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.578884 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.578930 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.578948 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.578973 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.578989 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:21Z","lastTransitionTime":"2025-11-22T04:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.681527 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.681585 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.681603 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.681627 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.681645 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:21Z","lastTransitionTime":"2025-11-22T04:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.784740 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.784834 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.784855 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.784878 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.784898 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:21Z","lastTransitionTime":"2025-11-22T04:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.888287 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.888347 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.888365 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.888392 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.888410 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:21Z","lastTransitionTime":"2025-11-22T04:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.991802 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.991871 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.991918 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.991939 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:21 crc kubenswrapper[4818]: I1122 04:49:21.991955 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:21Z","lastTransitionTime":"2025-11-22T04:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.094623 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.094695 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.094733 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.094763 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.094784 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:22Z","lastTransitionTime":"2025-11-22T04:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.198740 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.198813 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.198834 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.198863 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.198885 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:22Z","lastTransitionTime":"2025-11-22T04:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.290875 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:22 crc kubenswrapper[4818]: E1122 04:49:22.291054 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.291160 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:22 crc kubenswrapper[4818]: E1122 04:49:22.291466 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.301636 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.301707 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.301725 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.301745 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.301763 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:22Z","lastTransitionTime":"2025-11-22T04:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.404616 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.404717 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.404777 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.404812 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.404833 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:22Z","lastTransitionTime":"2025-11-22T04:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.508622 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.508699 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.508720 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.508749 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.508767 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:22Z","lastTransitionTime":"2025-11-22T04:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.612341 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.612398 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.612417 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.612440 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.612458 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:22Z","lastTransitionTime":"2025-11-22T04:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.715864 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.715922 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.715938 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.715960 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.715982 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:22Z","lastTransitionTime":"2025-11-22T04:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.819025 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.819115 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.819132 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.819157 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.819176 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:22Z","lastTransitionTime":"2025-11-22T04:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.921789 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.921850 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.921867 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.921891 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:22 crc kubenswrapper[4818]: I1122 04:49:22.921912 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:22Z","lastTransitionTime":"2025-11-22T04:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.024458 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.024519 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.024537 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.024561 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.024582 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:23Z","lastTransitionTime":"2025-11-22T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.128512 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.128592 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.128618 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.128649 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.128673 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:23Z","lastTransitionTime":"2025-11-22T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.231315 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.231428 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.231449 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.231528 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.231552 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:23Z","lastTransitionTime":"2025-11-22T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.290307 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.290451 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:23 crc kubenswrapper[4818]: E1122 04:49:23.290536 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:23 crc kubenswrapper[4818]: E1122 04:49:23.290868 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.291903 4818 scope.go:117] "RemoveContainer" containerID="37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba" Nov 22 04:49:23 crc kubenswrapper[4818]: E1122 04:49:23.292166 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p8s7r_openshift-ovn-kubernetes(f8f88a0c-c404-4e82-88c5-3e0f4f2b9025)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.335373 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.335447 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.335459 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.335479 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.335494 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:23Z","lastTransitionTime":"2025-11-22T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.418934 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.419032 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.419051 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.419075 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.419096 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:23Z","lastTransitionTime":"2025-11-22T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:23 crc kubenswrapper[4818]: E1122 04:49:23.441246 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.452691 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.452750 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.452770 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.452801 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.452824 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:23Z","lastTransitionTime":"2025-11-22T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:23 crc kubenswrapper[4818]: E1122 04:49:23.474030 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.480180 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.480509 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.480673 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.480813 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.480945 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:23Z","lastTransitionTime":"2025-11-22T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:23 crc kubenswrapper[4818]: E1122 04:49:23.503156 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.508027 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.508079 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.508103 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.508212 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.508244 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:23Z","lastTransitionTime":"2025-11-22T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:23 crc kubenswrapper[4818]: E1122 04:49:23.528948 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.534498 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.534722 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.534977 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.535188 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.535441 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:23Z","lastTransitionTime":"2025-11-22T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:23 crc kubenswrapper[4818]: E1122 04:49:23.555540 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T04:49:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"64c9fbbd-60c5-409c-a987-08eb342f18a6\\\",\\\"systemUUID\\\":\\\"e865ab18-bc50-41fa-b6be-f0543f192f64\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T04:49:23Z is after 2025-08-24T17:21:41Z" Nov 22 04:49:23 crc kubenswrapper[4818]: E1122 04:49:23.555768 4818 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.558080 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.558122 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.558139 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.558159 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.558176 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:23Z","lastTransitionTime":"2025-11-22T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.661494 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.661544 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.661562 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.661585 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.661603 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:23Z","lastTransitionTime":"2025-11-22T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.764337 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.764397 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.764424 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.764452 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.764472 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:23Z","lastTransitionTime":"2025-11-22T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.867298 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.867348 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.867364 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.867385 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.867428 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:23Z","lastTransitionTime":"2025-11-22T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.970041 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.970105 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.970122 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.970145 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:23 crc kubenswrapper[4818]: I1122 04:49:23.970165 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:23Z","lastTransitionTime":"2025-11-22T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.073153 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.073215 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.073233 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.073296 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.073317 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:24Z","lastTransitionTime":"2025-11-22T04:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.177043 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.177119 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.177137 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.177161 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.177178 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:24Z","lastTransitionTime":"2025-11-22T04:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.280534 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.280623 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.280648 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.280677 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.280702 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:24Z","lastTransitionTime":"2025-11-22T04:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.290926 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.291034 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:24 crc kubenswrapper[4818]: E1122 04:49:24.291140 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:24 crc kubenswrapper[4818]: E1122 04:49:24.291344 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.383665 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.383727 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.383745 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.383823 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.383842 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:24Z","lastTransitionTime":"2025-11-22T04:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.487075 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.487138 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.487160 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.487189 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.487213 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:24Z","lastTransitionTime":"2025-11-22T04:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.591219 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.591318 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.591338 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.591369 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.591398 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:24Z","lastTransitionTime":"2025-11-22T04:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.695244 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.695328 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.695344 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.695367 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.695386 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:24Z","lastTransitionTime":"2025-11-22T04:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.798993 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.799072 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.799091 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.799115 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.799135 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:24Z","lastTransitionTime":"2025-11-22T04:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.902428 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.902481 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.902498 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.902520 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:24 crc kubenswrapper[4818]: I1122 04:49:24.902539 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:24Z","lastTransitionTime":"2025-11-22T04:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.005315 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.005390 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.005415 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.005443 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.005464 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:25Z","lastTransitionTime":"2025-11-22T04:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.108644 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.108885 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.108904 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.108929 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.108946 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:25Z","lastTransitionTime":"2025-11-22T04:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.212300 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.212379 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.212407 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.212431 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.212450 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:25Z","lastTransitionTime":"2025-11-22T04:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.290693 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.290736 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:25 crc kubenswrapper[4818]: E1122 04:49:25.290883 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:25 crc kubenswrapper[4818]: E1122 04:49:25.291007 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.315287 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.315475 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.315631 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.316131 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.316190 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:25Z","lastTransitionTime":"2025-11-22T04:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.420196 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.420290 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.420312 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.420337 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.420355 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:25Z","lastTransitionTime":"2025-11-22T04:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.523424 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.523490 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.523509 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.523531 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.523548 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:25Z","lastTransitionTime":"2025-11-22T04:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.626336 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.626417 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.626436 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.626460 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.626478 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:25Z","lastTransitionTime":"2025-11-22T04:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.730122 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.730555 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.730726 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.730866 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.730985 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:25Z","lastTransitionTime":"2025-11-22T04:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.834225 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.834323 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.834342 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.834365 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.834382 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:25Z","lastTransitionTime":"2025-11-22T04:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.937685 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.938057 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.938198 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.938411 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:25 crc kubenswrapper[4818]: I1122 04:49:25.938555 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:25Z","lastTransitionTime":"2025-11-22T04:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.041431 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.041498 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.041520 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.041546 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.041565 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:26Z","lastTransitionTime":"2025-11-22T04:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.144929 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.145017 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.145038 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.145103 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.145121 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:26Z","lastTransitionTime":"2025-11-22T04:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.248413 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.248493 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.248531 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.248568 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.248609 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:26Z","lastTransitionTime":"2025-11-22T04:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.291437 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.291573 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:26 crc kubenswrapper[4818]: E1122 04:49:26.291746 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:26 crc kubenswrapper[4818]: E1122 04:49:26.291811 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.352169 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.352226 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.352250 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.352320 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.352343 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:26Z","lastTransitionTime":"2025-11-22T04:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.455233 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.455349 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.455410 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.455440 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.455463 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:26Z","lastTransitionTime":"2025-11-22T04:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.559625 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.559714 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.559731 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.559754 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.559772 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:26Z","lastTransitionTime":"2025-11-22T04:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.665240 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.665351 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.665374 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.665403 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.665424 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:26Z","lastTransitionTime":"2025-11-22T04:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.768713 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.768763 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.768780 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.768803 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.768820 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:26Z","lastTransitionTime":"2025-11-22T04:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.871391 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.871429 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.871438 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.871453 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.871464 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:26Z","lastTransitionTime":"2025-11-22T04:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.974059 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.974121 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.974144 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.974172 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:26 crc kubenswrapper[4818]: I1122 04:49:26.974193 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:26Z","lastTransitionTime":"2025-11-22T04:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.077318 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.077378 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.077401 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.077433 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.077454 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:27Z","lastTransitionTime":"2025-11-22T04:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.179988 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.180043 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.180062 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.180084 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.180104 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:27Z","lastTransitionTime":"2025-11-22T04:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.283609 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.283689 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.283713 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.283745 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.283767 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:27Z","lastTransitionTime":"2025-11-22T04:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.290085 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.290137 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:27 crc kubenswrapper[4818]: E1122 04:49:27.290237 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:27 crc kubenswrapper[4818]: E1122 04:49:27.290415 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.387209 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.387338 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.387370 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.387400 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.387422 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:27Z","lastTransitionTime":"2025-11-22T04:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.490771 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.490828 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.490846 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.490869 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.490886 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:27Z","lastTransitionTime":"2025-11-22T04:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.594904 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.594954 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.594970 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.594991 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.595007 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:27Z","lastTransitionTime":"2025-11-22T04:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.697196 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.697295 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.697320 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.697344 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.697361 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:27Z","lastTransitionTime":"2025-11-22T04:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.800564 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.800615 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.800628 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.800647 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.800660 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:27Z","lastTransitionTime":"2025-11-22T04:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.903439 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.903490 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.903509 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.903530 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:27 crc kubenswrapper[4818]: I1122 04:49:27.903548 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:27Z","lastTransitionTime":"2025-11-22T04:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.007062 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.007129 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.007145 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.007169 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.007191 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:28Z","lastTransitionTime":"2025-11-22T04:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.110175 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.110238 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.110284 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.110313 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.110332 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:28Z","lastTransitionTime":"2025-11-22T04:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:28 crc kubenswrapper[4818]: E1122 04:49:28.211303 4818 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.290884 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:28 crc kubenswrapper[4818]: E1122 04:49:28.291020 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.291023 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:28 crc kubenswrapper[4818]: E1122 04:49:28.291174 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.326538 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=31.32650883 podStartE2EDuration="31.32650883s" podCreationTimestamp="2025-11-22 04:48:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:28.311581802 +0000 UTC m=+120.885998369" watchObservedRunningTime="2025-11-22 04:49:28.32650883 +0000 UTC m=+120.900925407" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.387229 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-j9r6v" podStartSLOduration=97.387206847 podStartE2EDuration="1m37.387206847s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:28.386433007 +0000 UTC m=+120.960849584" watchObservedRunningTime="2025-11-22 04:49:28.387206847 +0000 UTC m=+120.961623384" Nov 22 04:49:28 crc kubenswrapper[4818]: E1122 04:49:28.427902 4818 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.445653 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-ct4xj" podStartSLOduration=97.445620275 podStartE2EDuration="1m37.445620275s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:28.443971832 +0000 UTC m=+121.018388379" watchObservedRunningTime="2025-11-22 04:49:28.445620275 +0000 UTC m=+121.020036802" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.445922 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=29.445916722 podStartE2EDuration="29.445916722s" podCreationTimestamp="2025-11-22 04:48:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:28.41890099 +0000 UTC m=+120.993317567" watchObservedRunningTime="2025-11-22 04:49:28.445916722 +0000 UTC m=+121.020333249" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.461713 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x52mk" podStartSLOduration=96.461695253 podStartE2EDuration="1m36.461695253s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:28.459748372 +0000 UTC m=+121.034164899" watchObservedRunningTime="2025-11-22 04:49:28.461695253 +0000 UTC m=+121.036111790" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.501561 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=23.501540068 podStartE2EDuration="23.501540068s" podCreationTimestamp="2025-11-22 04:49:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:28.501308642 +0000 UTC m=+121.075725189" watchObservedRunningTime="2025-11-22 04:49:28.501540068 +0000 UTC m=+121.075956595" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.501700 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=90.501693152 podStartE2EDuration="1m30.501693152s" podCreationTimestamp="2025-11-22 04:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:28.48241721 +0000 UTC m=+121.056833747" watchObservedRunningTime="2025-11-22 04:49:28.501693152 +0000 UTC m=+121.076109679" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.516754 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-x865d" podStartSLOduration=97.516735343 podStartE2EDuration="1m37.516735343s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:28.516266761 +0000 UTC m=+121.090683288" watchObservedRunningTime="2025-11-22 04:49:28.516735343 +0000 UTC m=+121.091151870" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.536581 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podStartSLOduration=97.536555677 podStartE2EDuration="1m37.536555677s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:28.52858121 +0000 UTC m=+121.102997737" watchObservedRunningTime="2025-11-22 04:49:28.536555677 +0000 UTC m=+121.110972224" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.537473 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-n4wst" podStartSLOduration=97.537462851 podStartE2EDuration="1m37.537462851s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:28.536747772 +0000 UTC m=+121.111164299" watchObservedRunningTime="2025-11-22 04:49:28.537462851 +0000 UTC m=+121.111879388" Nov 22 04:49:28 crc kubenswrapper[4818]: I1122 04:49:28.560350 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=65.560328915 podStartE2EDuration="1m5.560328915s" podCreationTimestamp="2025-11-22 04:48:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:28.559663018 +0000 UTC m=+121.134079545" watchObservedRunningTime="2025-11-22 04:49:28.560328915 +0000 UTC m=+121.134745452" Nov 22 04:49:29 crc kubenswrapper[4818]: I1122 04:49:29.031635 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x865d_d96879f8-2766-4f5e-bc3e-bbf8e5394a94/kube-multus/1.log" Nov 22 04:49:29 crc kubenswrapper[4818]: I1122 04:49:29.032476 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x865d_d96879f8-2766-4f5e-bc3e-bbf8e5394a94/kube-multus/0.log" Nov 22 04:49:29 crc kubenswrapper[4818]: I1122 04:49:29.032558 4818 generic.go:334] "Generic (PLEG): container finished" podID="d96879f8-2766-4f5e-bc3e-bbf8e5394a94" containerID="52d352c1148218337557f43c9772c5cc9866398ef20d8b5eb91eeff53c4cecbd" exitCode=1 Nov 22 04:49:29 crc kubenswrapper[4818]: I1122 04:49:29.032623 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x865d" event={"ID":"d96879f8-2766-4f5e-bc3e-bbf8e5394a94","Type":"ContainerDied","Data":"52d352c1148218337557f43c9772c5cc9866398ef20d8b5eb91eeff53c4cecbd"} Nov 22 04:49:29 crc kubenswrapper[4818]: I1122 04:49:29.032726 4818 scope.go:117] "RemoveContainer" containerID="27d8a9228da6feb61780c8066fe143e83d8d0467954bbb9c3857001ab21462c2" Nov 22 04:49:29 crc kubenswrapper[4818]: I1122 04:49:29.034204 4818 scope.go:117] "RemoveContainer" containerID="52d352c1148218337557f43c9772c5cc9866398ef20d8b5eb91eeff53c4cecbd" Nov 22 04:49:29 crc kubenswrapper[4818]: E1122 04:49:29.035005 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-x865d_openshift-multus(d96879f8-2766-4f5e-bc3e-bbf8e5394a94)\"" pod="openshift-multus/multus-x865d" podUID="d96879f8-2766-4f5e-bc3e-bbf8e5394a94" Nov 22 04:49:29 crc kubenswrapper[4818]: I1122 04:49:29.290159 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:29 crc kubenswrapper[4818]: I1122 04:49:29.290186 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:29 crc kubenswrapper[4818]: E1122 04:49:29.290411 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:29 crc kubenswrapper[4818]: E1122 04:49:29.290624 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:30 crc kubenswrapper[4818]: I1122 04:49:30.040157 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x865d_d96879f8-2766-4f5e-bc3e-bbf8e5394a94/kube-multus/1.log" Nov 22 04:49:30 crc kubenswrapper[4818]: I1122 04:49:30.291070 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:30 crc kubenswrapper[4818]: I1122 04:49:30.291148 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:30 crc kubenswrapper[4818]: E1122 04:49:30.291367 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:30 crc kubenswrapper[4818]: E1122 04:49:30.291533 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:31 crc kubenswrapper[4818]: I1122 04:49:31.290488 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:31 crc kubenswrapper[4818]: I1122 04:49:31.290506 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:31 crc kubenswrapper[4818]: E1122 04:49:31.290677 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:31 crc kubenswrapper[4818]: E1122 04:49:31.290811 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:32 crc kubenswrapper[4818]: I1122 04:49:32.290709 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:32 crc kubenswrapper[4818]: I1122 04:49:32.290724 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:32 crc kubenswrapper[4818]: E1122 04:49:32.290985 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:32 crc kubenswrapper[4818]: E1122 04:49:32.291134 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.290747 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.290759 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:33 crc kubenswrapper[4818]: E1122 04:49:33.290968 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:33 crc kubenswrapper[4818]: E1122 04:49:33.291095 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:33 crc kubenswrapper[4818]: E1122 04:49:33.429380 4818 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.764172 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.764205 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.764213 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.764227 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.764236 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T04:49:33Z","lastTransitionTime":"2025-11-22T04:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.817594 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5"] Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.818338 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.821463 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.822545 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.822812 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.823026 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.836977 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/5332e15b-8e38-4c7f-b762-4b3d71ffbaa5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-n62c5\" (UID: \"5332e15b-8e38-4c7f-b762-4b3d71ffbaa5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.837167 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5332e15b-8e38-4c7f-b762-4b3d71ffbaa5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-n62c5\" (UID: \"5332e15b-8e38-4c7f-b762-4b3d71ffbaa5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.837309 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5332e15b-8e38-4c7f-b762-4b3d71ffbaa5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-n62c5\" (UID: \"5332e15b-8e38-4c7f-b762-4b3d71ffbaa5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.837354 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/5332e15b-8e38-4c7f-b762-4b3d71ffbaa5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-n62c5\" (UID: \"5332e15b-8e38-4c7f-b762-4b3d71ffbaa5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.837398 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5332e15b-8e38-4c7f-b762-4b3d71ffbaa5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-n62c5\" (UID: \"5332e15b-8e38-4c7f-b762-4b3d71ffbaa5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.938204 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5332e15b-8e38-4c7f-b762-4b3d71ffbaa5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-n62c5\" (UID: \"5332e15b-8e38-4c7f-b762-4b3d71ffbaa5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.938337 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5332e15b-8e38-4c7f-b762-4b3d71ffbaa5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-n62c5\" (UID: \"5332e15b-8e38-4c7f-b762-4b3d71ffbaa5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.938365 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/5332e15b-8e38-4c7f-b762-4b3d71ffbaa5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-n62c5\" (UID: \"5332e15b-8e38-4c7f-b762-4b3d71ffbaa5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.938391 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5332e15b-8e38-4c7f-b762-4b3d71ffbaa5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-n62c5\" (UID: \"5332e15b-8e38-4c7f-b762-4b3d71ffbaa5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.939115 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/5332e15b-8e38-4c7f-b762-4b3d71ffbaa5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-n62c5\" (UID: \"5332e15b-8e38-4c7f-b762-4b3d71ffbaa5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.939165 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/5332e15b-8e38-4c7f-b762-4b3d71ffbaa5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-n62c5\" (UID: \"5332e15b-8e38-4c7f-b762-4b3d71ffbaa5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.938459 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/5332e15b-8e38-4c7f-b762-4b3d71ffbaa5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-n62c5\" (UID: \"5332e15b-8e38-4c7f-b762-4b3d71ffbaa5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.939234 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5332e15b-8e38-4c7f-b762-4b3d71ffbaa5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-n62c5\" (UID: \"5332e15b-8e38-4c7f-b762-4b3d71ffbaa5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.944064 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5332e15b-8e38-4c7f-b762-4b3d71ffbaa5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-n62c5\" (UID: \"5332e15b-8e38-4c7f-b762-4b3d71ffbaa5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5" Nov 22 04:49:33 crc kubenswrapper[4818]: I1122 04:49:33.957729 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5332e15b-8e38-4c7f-b762-4b3d71ffbaa5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-n62c5\" (UID: \"5332e15b-8e38-4c7f-b762-4b3d71ffbaa5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5" Nov 22 04:49:34 crc kubenswrapper[4818]: I1122 04:49:34.139591 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5" Nov 22 04:49:34 crc kubenswrapper[4818]: W1122 04:49:34.155850 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5332e15b_8e38_4c7f_b762_4b3d71ffbaa5.slice/crio-74a15ec3bc1ef96c3da29d8f9d04efdfd37f852efec06518e3b16ad575a0b15a WatchSource:0}: Error finding container 74a15ec3bc1ef96c3da29d8f9d04efdfd37f852efec06518e3b16ad575a0b15a: Status 404 returned error can't find the container with id 74a15ec3bc1ef96c3da29d8f9d04efdfd37f852efec06518e3b16ad575a0b15a Nov 22 04:49:34 crc kubenswrapper[4818]: I1122 04:49:34.290224 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:34 crc kubenswrapper[4818]: E1122 04:49:34.290457 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:34 crc kubenswrapper[4818]: I1122 04:49:34.290568 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:34 crc kubenswrapper[4818]: E1122 04:49:34.290695 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:35 crc kubenswrapper[4818]: I1122 04:49:35.060484 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5" event={"ID":"5332e15b-8e38-4c7f-b762-4b3d71ffbaa5","Type":"ContainerStarted","Data":"c2c07119b55c2abde083318243155699168c91807a82af2e98e99e78e607df54"} Nov 22 04:49:35 crc kubenswrapper[4818]: I1122 04:49:35.060556 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5" event={"ID":"5332e15b-8e38-4c7f-b762-4b3d71ffbaa5","Type":"ContainerStarted","Data":"74a15ec3bc1ef96c3da29d8f9d04efdfd37f852efec06518e3b16ad575a0b15a"} Nov 22 04:49:35 crc kubenswrapper[4818]: I1122 04:49:35.085980 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n62c5" podStartSLOduration=104.085953828 podStartE2EDuration="1m44.085953828s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:35.083562892 +0000 UTC m=+127.657979479" watchObservedRunningTime="2025-11-22 04:49:35.085953828 +0000 UTC m=+127.660370395" Nov 22 04:49:35 crc kubenswrapper[4818]: I1122 04:49:35.290379 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:35 crc kubenswrapper[4818]: I1122 04:49:35.290439 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:35 crc kubenswrapper[4818]: E1122 04:49:35.290583 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:35 crc kubenswrapper[4818]: E1122 04:49:35.290747 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:36 crc kubenswrapper[4818]: I1122 04:49:36.290569 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:36 crc kubenswrapper[4818]: E1122 04:49:36.290708 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:36 crc kubenswrapper[4818]: I1122 04:49:36.290907 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:36 crc kubenswrapper[4818]: E1122 04:49:36.291182 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:37 crc kubenswrapper[4818]: I1122 04:49:37.290283 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:37 crc kubenswrapper[4818]: I1122 04:49:37.290306 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:37 crc kubenswrapper[4818]: E1122 04:49:37.290419 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:37 crc kubenswrapper[4818]: E1122 04:49:37.290491 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:38 crc kubenswrapper[4818]: I1122 04:49:38.290933 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:38 crc kubenswrapper[4818]: E1122 04:49:38.293691 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:38 crc kubenswrapper[4818]: I1122 04:49:38.293896 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:38 crc kubenswrapper[4818]: E1122 04:49:38.294455 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:38 crc kubenswrapper[4818]: I1122 04:49:38.294702 4818 scope.go:117] "RemoveContainer" containerID="37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba" Nov 22 04:49:38 crc kubenswrapper[4818]: E1122 04:49:38.430709 4818 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 04:49:39 crc kubenswrapper[4818]: I1122 04:49:39.076113 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p8s7r_f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/ovnkube-controller/3.log" Nov 22 04:49:39 crc kubenswrapper[4818]: I1122 04:49:39.078795 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerStarted","Data":"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d"} Nov 22 04:49:39 crc kubenswrapper[4818]: I1122 04:49:39.079273 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:49:39 crc kubenswrapper[4818]: I1122 04:49:39.290681 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:39 crc kubenswrapper[4818]: I1122 04:49:39.290742 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:39 crc kubenswrapper[4818]: E1122 04:49:39.290823 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:39 crc kubenswrapper[4818]: E1122 04:49:39.290977 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:39 crc kubenswrapper[4818]: I1122 04:49:39.488919 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" podStartSLOduration=108.488896961 podStartE2EDuration="1m48.488896961s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:39.114785835 +0000 UTC m=+131.689202392" watchObservedRunningTime="2025-11-22 04:49:39.488896961 +0000 UTC m=+132.063313528" Nov 22 04:49:39 crc kubenswrapper[4818]: I1122 04:49:39.490857 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-djd77"] Nov 22 04:49:39 crc kubenswrapper[4818]: I1122 04:49:39.491166 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:39 crc kubenswrapper[4818]: E1122 04:49:39.491514 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:40 crc kubenswrapper[4818]: I1122 04:49:40.290601 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:40 crc kubenswrapper[4818]: E1122 04:49:40.290732 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:41 crc kubenswrapper[4818]: I1122 04:49:41.291206 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:41 crc kubenswrapper[4818]: I1122 04:49:41.291216 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:41 crc kubenswrapper[4818]: E1122 04:49:41.291480 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:41 crc kubenswrapper[4818]: I1122 04:49:41.291234 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:41 crc kubenswrapper[4818]: E1122 04:49:41.291643 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:41 crc kubenswrapper[4818]: E1122 04:49:41.291920 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:42 crc kubenswrapper[4818]: I1122 04:49:42.290239 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:42 crc kubenswrapper[4818]: E1122 04:49:42.290460 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:43 crc kubenswrapper[4818]: I1122 04:49:43.291156 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:43 crc kubenswrapper[4818]: I1122 04:49:43.291236 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:43 crc kubenswrapper[4818]: I1122 04:49:43.291314 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:43 crc kubenswrapper[4818]: E1122 04:49:43.291406 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:43 crc kubenswrapper[4818]: E1122 04:49:43.291514 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:43 crc kubenswrapper[4818]: E1122 04:49:43.291631 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:43 crc kubenswrapper[4818]: E1122 04:49:43.433094 4818 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 04:49:44 crc kubenswrapper[4818]: I1122 04:49:44.290094 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:44 crc kubenswrapper[4818]: E1122 04:49:44.290432 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:44 crc kubenswrapper[4818]: I1122 04:49:44.290867 4818 scope.go:117] "RemoveContainer" containerID="52d352c1148218337557f43c9772c5cc9866398ef20d8b5eb91eeff53c4cecbd" Nov 22 04:49:45 crc kubenswrapper[4818]: I1122 04:49:45.104707 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x865d_d96879f8-2766-4f5e-bc3e-bbf8e5394a94/kube-multus/1.log" Nov 22 04:49:45 crc kubenswrapper[4818]: I1122 04:49:45.105069 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x865d" event={"ID":"d96879f8-2766-4f5e-bc3e-bbf8e5394a94","Type":"ContainerStarted","Data":"6b22db5b4662967112f000f6af50106e3510dc0314e236e402abb833ed2bad54"} Nov 22 04:49:45 crc kubenswrapper[4818]: I1122 04:49:45.290371 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:45 crc kubenswrapper[4818]: I1122 04:49:45.290528 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:45 crc kubenswrapper[4818]: I1122 04:49:45.290610 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:45 crc kubenswrapper[4818]: E1122 04:49:45.290559 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:45 crc kubenswrapper[4818]: E1122 04:49:45.290821 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:45 crc kubenswrapper[4818]: E1122 04:49:45.290948 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:46 crc kubenswrapper[4818]: I1122 04:49:46.291164 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:46 crc kubenswrapper[4818]: E1122 04:49:46.291411 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:47 crc kubenswrapper[4818]: I1122 04:49:47.290641 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:47 crc kubenswrapper[4818]: I1122 04:49:47.290677 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:47 crc kubenswrapper[4818]: I1122 04:49:47.290697 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:47 crc kubenswrapper[4818]: E1122 04:49:47.290814 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 04:49:47 crc kubenswrapper[4818]: E1122 04:49:47.290926 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 04:49:47 crc kubenswrapper[4818]: E1122 04:49:47.291033 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-djd77" podUID="429a37e7-cb52-4e76-89ee-16f2b9e6fbfe" Nov 22 04:49:48 crc kubenswrapper[4818]: I1122 04:49:48.290457 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:48 crc kubenswrapper[4818]: E1122 04:49:48.292289 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 04:49:49 crc kubenswrapper[4818]: I1122 04:49:49.290322 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:49:49 crc kubenswrapper[4818]: I1122 04:49:49.290358 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:49:49 crc kubenswrapper[4818]: I1122 04:49:49.290397 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:49:49 crc kubenswrapper[4818]: I1122 04:49:49.294338 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 22 04:49:49 crc kubenswrapper[4818]: I1122 04:49:49.294572 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 22 04:49:49 crc kubenswrapper[4818]: I1122 04:49:49.294780 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 22 04:49:49 crc kubenswrapper[4818]: I1122 04:49:49.294921 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 22 04:49:49 crc kubenswrapper[4818]: I1122 04:49:49.295306 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 22 04:49:49 crc kubenswrapper[4818]: I1122 04:49:49.295563 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 22 04:49:50 crc kubenswrapper[4818]: I1122 04:49:50.291214 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.266170 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.321697 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tjh65"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.322325 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.324609 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-8kzxl"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.325476 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-8kzxl" Nov 22 04:49:54 crc kubenswrapper[4818]: W1122 04:49:54.326558 4818 reflector.go:561] object-"openshift-controller-manager"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Nov 22 04:49:54 crc kubenswrapper[4818]: E1122 04:49:54.326631 4818 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 22 04:49:54 crc kubenswrapper[4818]: W1122 04:49:54.326641 4818 reflector.go:561] object-"openshift-controller-manager"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Nov 22 04:49:54 crc kubenswrapper[4818]: W1122 04:49:54.326723 4818 reflector.go:561] object-"openshift-controller-manager"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Nov 22 04:49:54 crc kubenswrapper[4818]: W1122 04:49:54.326741 4818 reflector.go:561] object-"openshift-controller-manager"/"client-ca": failed to list *v1.ConfigMap: configmaps "client-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Nov 22 04:49:54 crc kubenswrapper[4818]: E1122 04:49:54.326781 4818 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 22 04:49:54 crc kubenswrapper[4818]: E1122 04:49:54.326799 4818 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"client-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"client-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 22 04:49:54 crc kubenswrapper[4818]: E1122 04:49:54.326750 4818 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.328505 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.329325 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.349582 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-9nsph"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.351820 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.354223 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.354528 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.354672 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: W1122 04:49:54.355066 4818 reflector.go:561] object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c": failed to list *v1.Secret: secrets "openshift-controller-manager-sa-dockercfg-msq4c" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Nov 22 04:49:54 crc kubenswrapper[4818]: E1122 04:49:54.355120 4818 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"openshift-controller-manager-sa-dockercfg-msq4c\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-controller-manager-sa-dockercfg-msq4c\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 22 04:49:54 crc kubenswrapper[4818]: W1122 04:49:54.355211 4818 reflector.go:561] object-"openshift-machine-api"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Nov 22 04:49:54 crc kubenswrapper[4818]: E1122 04:49:54.355242 4818 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.355320 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" Nov 22 04:49:54 crc kubenswrapper[4818]: W1122 04:49:54.355355 4818 reflector.go:561] object-"openshift-controller-manager"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Nov 22 04:49:54 crc kubenswrapper[4818]: E1122 04:49:54.355378 4818 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.355473 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: W1122 04:49:54.356107 4818 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7": failed to list *v1.Secret: secrets "machine-api-operator-dockercfg-mfbb7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.356160 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 22 04:49:54 crc kubenswrapper[4818]: E1122 04:49:54.356179 4818 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-dockercfg-mfbb7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-api-operator-dockercfg-mfbb7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.357205 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.359138 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-h7srs"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.360452 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.370083 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.372876 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.373700 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.378492 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.378714 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.378789 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.378920 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.379027 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.379070 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.379134 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.380608 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.381852 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.382085 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.382182 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.382208 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j7ttg"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.382712 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mkgj4"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.382986 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-rcvlq"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.383301 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-s8xvt"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.383705 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-s8xvt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.384055 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h7srs" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.384317 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j7ttg" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.384430 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.384724 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-rcvlq" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.385134 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.385397 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.385575 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.385615 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.385684 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.385721 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.385883 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.386652 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.388325 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xvd5r"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.388908 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j7ttg"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.389019 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xvd5r" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.392637 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-9nsph"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.392675 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-44grv"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.393086 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-44grv" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.393474 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-dvt5z"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.397103 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.397303 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.397443 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.398029 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-dvt5z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.398435 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.398803 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.399404 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.399906 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-887w5"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.399999 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.400432 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-887w5" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.400583 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.401246 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.402219 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-8kzxl"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.405796 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ps29"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.406159 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-28f5t"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.402275 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.406453 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-28f5t" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.402309 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.406610 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ps29" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.403228 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.407051 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7f58c"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.403378 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.403641 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.407366 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7f58c" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.403682 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.403713 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.403714 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.403871 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.404016 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.404077 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.404325 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.404359 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.408041 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.408796 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-h56w2"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.408980 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.409174 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.409315 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.409393 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.409445 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.409526 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.409699 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-q6xbv"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.409952 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.409747 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.410155 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.410196 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q6xbv" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.410157 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.410627 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rsc75"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.429189 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.429316 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.429940 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.430000 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.430859 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.431808 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.432018 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.432290 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.432375 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.432338 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.432630 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.433161 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rr7gt"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.433805 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.433958 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.434091 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.434223 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.434369 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.434865 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.434917 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rr7gt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.435023 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.435287 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rsc75" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.435731 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.435886 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.436085 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.436190 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.435744 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.449627 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.453173 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.455985 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dhhh4"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.456617 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tjh65"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.456701 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.456710 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.457824 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-4pxtx"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.458641 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jtg8k"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.458680 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4pxtx" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.459194 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.459504 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jtg8k" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.459559 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.459611 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.459711 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.459834 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cd4qt"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.459994 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.460286 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.463018 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.463129 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.463779 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.463859 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.464205 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-rcvlq"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.465447 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nz99w"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.465814 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.466117 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.466816 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-5gfmz"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.467363 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.467614 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.467674 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.468094 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.469063 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.469593 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-jp5fs"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.470332 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fxwd8"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.470670 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-jp5fs" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.472993 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.473114 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fxwd8" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.474313 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.477571 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.479029 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.486305 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.494208 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.496645 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.498176 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7bmhw"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.498777 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7bmhw" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.500038 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-z7mtd"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.501224 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-z7mtd" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.508661 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7337e50a-daf8-4f50-9060-ec2a14752444-trusted-ca\") pod \"console-operator-58897d9998-44grv\" (UID: \"7337e50a-daf8-4f50-9060-ec2a14752444\") " pod="openshift-console-operator/console-operator-58897d9998-44grv" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.508707 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.508733 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1c05f6d6-e233-46ca-91df-eaba2614018f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-92bz6\" (UID: \"1c05f6d6-e233-46ca-91df-eaba2614018f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.508767 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqrct\" (UniqueName: \"kubernetes.io/projected/353ed841-0d3b-46df-beec-1906f6d801fd-kube-api-access-fqrct\") pod \"route-controller-manager-6576b87f9c-4lz4x\" (UID: \"353ed841-0d3b-46df-beec-1906f6d801fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.508790 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0dcaa80a-8993-4452-9acb-b8010eade7c6-audit-dir\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.508813 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/353ed841-0d3b-46df-beec-1906f6d801fd-config\") pod \"route-controller-manager-6576b87f9c-4lz4x\" (UID: \"353ed841-0d3b-46df-beec-1906f6d801fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.508837 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/091bb385-63c1-443f-9aec-bedd08514ee8-audit\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.508859 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/85928d04-1176-4b1d-8bb7-d55376576225-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2ps29\" (UID: \"85928d04-1176-4b1d-8bb7-d55376576225\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ps29" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.508882 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/968b0694-9a30-49ca-bee7-d1eadbb31505-machine-approver-tls\") pod \"machine-approver-56656f9798-h7srs\" (UID: \"968b0694-9a30-49ca-bee7-d1eadbb31505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h7srs" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.508903 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-audit-policies\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.508925 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-config\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.508946 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/670a3117-320e-477c-8c4b-a54fe5a4ae9c-console-oauth-config\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.508968 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/091bb385-63c1-443f-9aec-bedd08514ee8-config\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.508990 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtpcw\" (UniqueName: \"kubernetes.io/projected/4bfeb5de-dcfb-4f0b-b118-112ee3262c18-kube-api-access-qtpcw\") pod \"dns-operator-744455d44c-dvt5z\" (UID: \"4bfeb5de-dcfb-4f0b-b118-112ee3262c18\") " pod="openshift-dns-operator/dns-operator-744455d44c-dvt5z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.509014 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcll2\" (UniqueName: \"kubernetes.io/projected/8d3b850f-8917-46af-b713-891a7c979d2e-kube-api-access-tcll2\") pod \"machine-api-operator-5694c8668f-8kzxl\" (UID: \"8d3b850f-8917-46af-b713-891a7c979d2e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8kzxl" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.509167 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/10918dd9-6412-4172-a0f3-eac3fbd10da1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.509235 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/968b0694-9a30-49ca-bee7-d1eadbb31505-config\") pod \"machine-approver-56656f9798-h7srs\" (UID: \"968b0694-9a30-49ca-bee7-d1eadbb31505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h7srs" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.509292 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1c05f6d6-e233-46ca-91df-eaba2614018f-trusted-ca\") pod \"ingress-operator-5b745b69d9-92bz6\" (UID: \"1c05f6d6-e233-46ca-91df-eaba2614018f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.509318 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45ba8e80-12bf-472e-a0b7-b25601a9074e-config\") pod \"kube-controller-manager-operator-78b949d7b-887w5\" (UID: \"45ba8e80-12bf-472e-a0b7-b25601a9074e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-887w5" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.509338 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-service-ca\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.509359 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4bfeb5de-dcfb-4f0b-b118-112ee3262c18-metrics-tls\") pod \"dns-operator-744455d44c-dvt5z\" (UID: \"4bfeb5de-dcfb-4f0b-b118-112ee3262c18\") " pod="openshift-dns-operator/dns-operator-744455d44c-dvt5z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.509582 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zftfz"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.509568 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/10918dd9-6412-4172-a0f3-eac3fbd10da1-audit-policies\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.509643 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10918dd9-6412-4172-a0f3-eac3fbd10da1-serving-cert\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.509667 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5484670-b8e2-40b9-a36b-3fb47bb89421-serving-cert\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.509690 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/312c389c-57c0-4709-9320-777e4659d499-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-28f5t\" (UID: \"312c389c-57c0-4709-9320-777e4659d499\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-28f5t" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.509725 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/312c389c-57c0-4709-9320-777e4659d499-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-28f5t\" (UID: \"312c389c-57c0-4709-9320-777e4659d499\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-28f5t" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.509739 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.509760 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/85928d04-1176-4b1d-8bb7-d55376576225-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2ps29\" (UID: \"85928d04-1176-4b1d-8bb7-d55376576225\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ps29" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.509775 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8d3b850f-8917-46af-b713-891a7c979d2e-images\") pod \"machine-api-operator-5694c8668f-8kzxl\" (UID: \"8d3b850f-8917-46af-b713-891a7c979d2e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8kzxl" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.509838 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e48b2f0e-a02f-48c6-b513-33537b0108b6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xvd5r\" (UID: \"e48b2f0e-a02f-48c6-b513-33537b0108b6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xvd5r" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.509885 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gsdk\" (UniqueName: \"kubernetes.io/projected/1c05f6d6-e233-46ca-91df-eaba2614018f-kube-api-access-8gsdk\") pod \"ingress-operator-5b745b69d9-92bz6\" (UID: \"1c05f6d6-e233-46ca-91df-eaba2614018f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.509913 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq7dc\" (UniqueName: \"kubernetes.io/projected/091bb385-63c1-443f-9aec-bedd08514ee8-kube-api-access-jq7dc\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.509938 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md47q\" (UniqueName: \"kubernetes.io/projected/10918dd9-6412-4172-a0f3-eac3fbd10da1-kube-api-access-md47q\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.510553 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-console-config\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.510595 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdkln\" (UniqueName: \"kubernetes.io/projected/968b0694-9a30-49ca-bee7-d1eadbb31505-kube-api-access-jdkln\") pod \"machine-approver-56656f9798-h7srs\" (UID: \"968b0694-9a30-49ca-bee7-d1eadbb31505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h7srs" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.510648 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.510689 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23111038-ec66-4295-a6eb-b847c52a0c81-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7f58c\" (UID: \"23111038-ec66-4295-a6eb-b847c52a0c81\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7f58c" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.510737 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/091bb385-63c1-443f-9aec-bedd08514ee8-image-import-ca\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.510757 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvx6l\" (UniqueName: \"kubernetes.io/projected/85928d04-1176-4b1d-8bb7-d55376576225-kube-api-access-mvx6l\") pod \"cluster-image-registry-operator-dc59b4c8b-2ps29\" (UID: \"85928d04-1176-4b1d-8bb7-d55376576225\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ps29" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.510806 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgpx6\" (UniqueName: \"kubernetes.io/projected/5a66393b-1460-4ef7-9c8a-2001a0d187fb-kube-api-access-jgpx6\") pod \"authentication-operator-69f744f599-s8xvt\" (UID: \"5a66393b-1460-4ef7-9c8a-2001a0d187fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8xvt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.510833 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-oauth-serving-cert\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.510854 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e48b2f0e-a02f-48c6-b513-33537b0108b6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xvd5r\" (UID: \"e48b2f0e-a02f-48c6-b513-33537b0108b6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xvd5r" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.511105 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/312c389c-57c0-4709-9320-777e4659d499-config\") pod \"kube-apiserver-operator-766d6c64bb-28f5t\" (UID: \"312c389c-57c0-4709-9320-777e4659d499\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-28f5t" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.511127 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a66393b-1460-4ef7-9c8a-2001a0d187fb-service-ca-bundle\") pod \"authentication-operator-69f744f599-s8xvt\" (UID: \"5a66393b-1460-4ef7-9c8a-2001a0d187fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8xvt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.511175 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.511199 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23111038-ec66-4295-a6eb-b847c52a0c81-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7f58c\" (UID: \"23111038-ec66-4295-a6eb-b847c52a0c81\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7f58c" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.511238 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/45ba8e80-12bf-472e-a0b7-b25601a9074e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-887w5\" (UID: \"45ba8e80-12bf-472e-a0b7-b25601a9074e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-887w5" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.511279 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.511335 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zftfz" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.511331 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/85928d04-1176-4b1d-8bb7-d55376576225-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2ps29\" (UID: \"85928d04-1176-4b1d-8bb7-d55376576225\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ps29" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.511457 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a66393b-1460-4ef7-9c8a-2001a0d187fb-serving-cert\") pod \"authentication-operator-69f744f599-s8xvt\" (UID: \"5a66393b-1460-4ef7-9c8a-2001a0d187fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8xvt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.511510 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwpwz\" (UniqueName: \"kubernetes.io/projected/7337e50a-daf8-4f50-9060-ec2a14752444-kube-api-access-kwpwz\") pod \"console-operator-58897d9998-44grv\" (UID: \"7337e50a-daf8-4f50-9060-ec2a14752444\") " pod="openshift-console-operator/console-operator-58897d9998-44grv" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.511535 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1c05f6d6-e233-46ca-91df-eaba2614018f-metrics-tls\") pod \"ingress-operator-5b745b69d9-92bz6\" (UID: \"1c05f6d6-e233-46ca-91df-eaba2614018f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.511581 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9zfh\" (UniqueName: \"kubernetes.io/projected/d7c47e41-70c7-4784-aa37-c609650e3e1b-kube-api-access-h9zfh\") pod \"migrator-59844c95c7-q6xbv\" (UID: \"d7c47e41-70c7-4784-aa37-c609650e3e1b\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q6xbv" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.511605 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j7fh\" (UniqueName: \"kubernetes.io/projected/670a3117-320e-477c-8c4b-a54fe5a4ae9c-kube-api-access-8j7fh\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.511680 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95th8\" (UniqueName: \"kubernetes.io/projected/61feab0c-5f50-49fe-bf35-de2a5f66bf32-kube-api-access-95th8\") pod \"downloads-7954f5f757-rcvlq\" (UID: \"61feab0c-5f50-49fe-bf35-de2a5f66bf32\") " pod="openshift-console/downloads-7954f5f757-rcvlq" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.511707 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a66393b-1460-4ef7-9c8a-2001a0d187fb-config\") pod \"authentication-operator-69f744f599-s8xvt\" (UID: \"5a66393b-1460-4ef7-9c8a-2001a0d187fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8xvt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.511756 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58r46\" (UniqueName: \"kubernetes.io/projected/d5484670-b8e2-40b9-a36b-3fb47bb89421-kube-api-access-58r46\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.511777 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a66393b-1460-4ef7-9c8a-2001a0d187fb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-s8xvt\" (UID: \"5a66393b-1460-4ef7-9c8a-2001a0d187fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8xvt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.511827 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.511852 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.511877 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87hd4\" (UniqueName: \"kubernetes.io/projected/e48b2f0e-a02f-48c6-b513-33537b0108b6-kube-api-access-87hd4\") pod \"openshift-apiserver-operator-796bbdcf4f-xvd5r\" (UID: \"e48b2f0e-a02f-48c6-b513-33537b0108b6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xvd5r" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512012 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/091bb385-63c1-443f-9aec-bedd08514ee8-encryption-config\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512040 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/10918dd9-6412-4172-a0f3-eac3fbd10da1-etcd-client\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512106 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23111038-ec66-4295-a6eb-b847c52a0c81-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7f58c\" (UID: \"23111038-ec66-4295-a6eb-b847c52a0c81\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7f58c" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512155 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/353ed841-0d3b-46df-beec-1906f6d801fd-serving-cert\") pod \"route-controller-manager-6576b87f9c-4lz4x\" (UID: \"353ed841-0d3b-46df-beec-1906f6d801fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512184 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/10918dd9-6412-4172-a0f3-eac3fbd10da1-audit-dir\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512208 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512286 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/091bb385-63c1-443f-9aec-bedd08514ee8-trusted-ca-bundle\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512310 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjkhp\" (UniqueName: \"kubernetes.io/projected/0dcaa80a-8993-4452-9acb-b8010eade7c6-kube-api-access-sjkhp\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512361 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512396 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/091bb385-63c1-443f-9aec-bedd08514ee8-node-pullsecrets\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512438 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/091bb385-63c1-443f-9aec-bedd08514ee8-serving-cert\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512461 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45ba8e80-12bf-472e-a0b7-b25601a9074e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-887w5\" (UID: \"45ba8e80-12bf-472e-a0b7-b25601a9074e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-887w5" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512539 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-trusted-ca-bundle\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512572 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8d3b850f-8917-46af-b713-891a7c979d2e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-8kzxl\" (UID: \"8d3b850f-8917-46af-b713-891a7c979d2e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8kzxl" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512630 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/353ed841-0d3b-46df-beec-1906f6d801fd-client-ca\") pod \"route-controller-manager-6576b87f9c-4lz4x\" (UID: \"353ed841-0d3b-46df-beec-1906f6d801fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512660 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/091bb385-63c1-443f-9aec-bedd08514ee8-audit-dir\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512711 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/968b0694-9a30-49ca-bee7-d1eadbb31505-auth-proxy-config\") pod \"machine-approver-56656f9798-h7srs\" (UID: \"968b0694-9a30-49ca-bee7-d1eadbb31505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h7srs" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512744 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/62bdffb2-a1c7-4960-81aa-b570834f78cd-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-j7ttg\" (UID: \"62bdffb2-a1c7-4960-81aa-b570834f78cd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j7ttg" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512811 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512840 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512866 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/091bb385-63c1-443f-9aec-bedd08514ee8-etcd-serving-ca\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.512889 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs2s8\" (UniqueName: \"kubernetes.io/projected/62bdffb2-a1c7-4960-81aa-b570834f78cd-kube-api-access-bs2s8\") pod \"cluster-samples-operator-665b6dd947-j7ttg\" (UID: \"62bdffb2-a1c7-4960-81aa-b570834f78cd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j7ttg" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.513189 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7337e50a-daf8-4f50-9060-ec2a14752444-serving-cert\") pod \"console-operator-58897d9998-44grv\" (UID: \"7337e50a-daf8-4f50-9060-ec2a14752444\") " pod="openshift-console-operator/console-operator-58897d9998-44grv" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.513206 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/10918dd9-6412-4172-a0f3-eac3fbd10da1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.513226 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/10918dd9-6412-4172-a0f3-eac3fbd10da1-encryption-config\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.513244 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-client-ca\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.513274 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d3b850f-8917-46af-b713-891a7c979d2e-config\") pod \"machine-api-operator-5694c8668f-8kzxl\" (UID: \"8d3b850f-8917-46af-b713-891a7c979d2e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8kzxl" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.513293 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.513313 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/091bb385-63c1-443f-9aec-bedd08514ee8-etcd-client\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.513329 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/670a3117-320e-477c-8c4b-a54fe5a4ae9c-console-serving-cert\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.513347 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7337e50a-daf8-4f50-9060-ec2a14752444-config\") pod \"console-operator-58897d9998-44grv\" (UID: \"7337e50a-daf8-4f50-9060-ec2a14752444\") " pod="openshift-console-operator/console-operator-58897d9998-44grv" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.513930 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.515282 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.516890 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqnns"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.517001 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.518181 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xvd5r"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.518273 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqnns" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.519532 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.521248 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-44grv"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.521383 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.522399 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-28f5t"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.523402 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-s8xvt"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.524336 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jtg8k"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.525292 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cd4qt"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.526638 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-ccd5x"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.527076 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.528317 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-ccd5x" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.535288 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-8zp8h"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.536679 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7f58c"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.536786 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.539509 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fxwd8"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.542183 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mkgj4"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.544314 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.544622 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-dvt5z"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.546440 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zftfz"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.548053 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.551163 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rr7gt"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.552292 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-ccd5x"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.553734 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.555268 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nz99w"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.556273 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-jp5fs"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.557810 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.558840 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-887w5"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.560077 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-h56w2"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.561136 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-4pxtx"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.562372 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rsc75"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.563766 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.564467 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ps29"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.566037 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.567389 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.568793 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7bmhw"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.570676 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-8zp8h"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.571122 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.572153 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dhhh4"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.573501 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-q6xbv"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.574672 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqnns"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.576011 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.582096 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-b86wj"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.583628 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-b86wj" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.584141 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-b86wj"] Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.584598 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.604362 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.614478 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3e2ce5bb-1abf-47c6-b6d2-59389118cff0-profile-collector-cert\") pod \"catalog-operator-68c6474976-btgsr\" (UID: \"3e2ce5bb-1abf-47c6-b6d2-59389118cff0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.614517 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9zfh\" (UniqueName: \"kubernetes.io/projected/d7c47e41-70c7-4784-aa37-c609650e3e1b-kube-api-access-h9zfh\") pod \"migrator-59844c95c7-q6xbv\" (UID: \"d7c47e41-70c7-4784-aa37-c609650e3e1b\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q6xbv" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.614542 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58r46\" (UniqueName: \"kubernetes.io/projected/d5484670-b8e2-40b9-a36b-3fb47bb89421-kube-api-access-58r46\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.614569 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzppq\" (UniqueName: \"kubernetes.io/projected/751d9838-4c34-4251-b3a6-1d808f98a034-kube-api-access-xzppq\") pod \"collect-profiles-29396445-v6dn4\" (UID: \"751d9838-4c34-4251-b3a6-1d808f98a034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.614592 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/091bb385-63c1-443f-9aec-bedd08514ee8-encryption-config\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.614615 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87hd4\" (UniqueName: \"kubernetes.io/projected/e48b2f0e-a02f-48c6-b513-33537b0108b6-kube-api-access-87hd4\") pod \"openshift-apiserver-operator-796bbdcf4f-xvd5r\" (UID: \"e48b2f0e-a02f-48c6-b513-33537b0108b6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xvd5r" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.614652 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/eb0cffc3-bf07-411c-822e-aa45c58387cb-default-certificate\") pod \"router-default-5444994796-5gfmz\" (UID: \"eb0cffc3-bf07-411c-822e-aa45c58387cb\") " pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.614790 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/353ed841-0d3b-46df-beec-1906f6d801fd-serving-cert\") pod \"route-controller-manager-6576b87f9c-4lz4x\" (UID: \"353ed841-0d3b-46df-beec-1906f6d801fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.614867 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/091bb385-63c1-443f-9aec-bedd08514ee8-trusted-ca-bundle\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.615721 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjkhp\" (UniqueName: \"kubernetes.io/projected/0dcaa80a-8993-4452-9acb-b8010eade7c6-kube-api-access-sjkhp\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.615773 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8d3b850f-8917-46af-b713-891a7c979d2e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-8kzxl\" (UID: \"8d3b850f-8917-46af-b713-891a7c979d2e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8kzxl" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.615794 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.615813 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/091bb385-63c1-443f-9aec-bedd08514ee8-node-pullsecrets\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.615823 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/091bb385-63c1-443f-9aec-bedd08514ee8-trusted-ca-bundle\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.615829 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45ba8e80-12bf-472e-a0b7-b25601a9074e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-887w5\" (UID: \"45ba8e80-12bf-472e-a0b7-b25601a9074e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-887w5" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.615879 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-trusted-ca-bundle\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.615907 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk6gh\" (UniqueName: \"kubernetes.io/projected/a33da900-6b7c-4157-8a6e-fde1cddf51ca-kube-api-access-zk6gh\") pod \"csi-hostpathplugin-8zp8h\" (UID: \"a33da900-6b7c-4157-8a6e-fde1cddf51ca\") " pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.615934 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/091bb385-63c1-443f-9aec-bedd08514ee8-audit-dir\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.615959 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/968b0694-9a30-49ca-bee7-d1eadbb31505-auth-proxy-config\") pod \"machine-approver-56656f9798-h7srs\" (UID: \"968b0694-9a30-49ca-bee7-d1eadbb31505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h7srs" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.615983 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpq48\" (UniqueName: \"kubernetes.io/projected/1d4ef859-1409-451a-a2e1-4c1c8c6d473b-kube-api-access-kpq48\") pod \"marketplace-operator-79b997595-nz99w\" (UID: \"1d4ef859-1409-451a-a2e1-4c1c8c6d473b\") " pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616005 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a33da900-6b7c-4157-8a6e-fde1cddf51ca-socket-dir\") pod \"csi-hostpathplugin-8zp8h\" (UID: \"a33da900-6b7c-4157-8a6e-fde1cddf51ca\") " pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616026 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2a1ecdb-4740-43c4-ba73-19d1f15465e1-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rsc75\" (UID: \"f2a1ecdb-4740-43c4-ba73-19d1f15465e1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rsc75" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616047 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/40771f13-f771-4e07-8ad9-668a3e48fb3b-etcd-client\") pod \"etcd-operator-b45778765-dhhh4\" (UID: \"40771f13-f771-4e07-8ad9-668a3e48fb3b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616080 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616106 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616129 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/091bb385-63c1-443f-9aec-bedd08514ee8-etcd-serving-ca\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616154 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs2s8\" (UniqueName: \"kubernetes.io/projected/62bdffb2-a1c7-4960-81aa-b570834f78cd-kube-api-access-bs2s8\") pod \"cluster-samples-operator-665b6dd947-j7ttg\" (UID: \"62bdffb2-a1c7-4960-81aa-b570834f78cd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j7ttg" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616178 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/751d9838-4c34-4251-b3a6-1d808f98a034-secret-volume\") pod \"collect-profiles-29396445-v6dn4\" (UID: \"751d9838-4c34-4251-b3a6-1d808f98a034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616200 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a-config\") pod \"service-ca-operator-777779d784-fxwd8\" (UID: \"b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fxwd8" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616226 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-client-ca\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616263 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/091bb385-63c1-443f-9aec-bedd08514ee8-etcd-client\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616292 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616319 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cccc9\" (UniqueName: \"kubernetes.io/projected/f2a1ecdb-4740-43c4-ba73-19d1f15465e1-kube-api-access-cccc9\") pod \"openshift-controller-manager-operator-756b6f6bc6-rsc75\" (UID: \"f2a1ecdb-4740-43c4-ba73-19d1f15465e1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rsc75" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616340 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40771f13-f771-4e07-8ad9-668a3e48fb3b-config\") pod \"etcd-operator-b45778765-dhhh4\" (UID: \"40771f13-f771-4e07-8ad9-668a3e48fb3b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616366 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7337e50a-daf8-4f50-9060-ec2a14752444-trusted-ca\") pod \"console-operator-58897d9998-44grv\" (UID: \"7337e50a-daf8-4f50-9060-ec2a14752444\") " pod="openshift-console-operator/console-operator-58897d9998-44grv" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616391 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqrct\" (UniqueName: \"kubernetes.io/projected/353ed841-0d3b-46df-beec-1906f6d801fd-kube-api-access-fqrct\") pod \"route-controller-manager-6576b87f9c-4lz4x\" (UID: \"353ed841-0d3b-46df-beec-1906f6d801fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616413 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-audit-policies\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616434 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/40771f13-f771-4e07-8ad9-668a3e48fb3b-etcd-ca\") pod \"etcd-operator-b45778765-dhhh4\" (UID: \"40771f13-f771-4e07-8ad9-668a3e48fb3b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616456 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-4pxtx\" (UID: \"54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4pxtx" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616693 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/eb0cffc3-bf07-411c-822e-aa45c58387cb-stats-auth\") pod \"router-default-5444994796-5gfmz\" (UID: \"eb0cffc3-bf07-411c-822e-aa45c58387cb\") " pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616756 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/353ed841-0d3b-46df-beec-1906f6d801fd-config\") pod \"route-controller-manager-6576b87f9c-4lz4x\" (UID: \"353ed841-0d3b-46df-beec-1906f6d801fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616787 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/091bb385-63c1-443f-9aec-bedd08514ee8-audit\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616842 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/85928d04-1176-4b1d-8bb7-d55376576225-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2ps29\" (UID: \"85928d04-1176-4b1d-8bb7-d55376576225\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ps29" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616872 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7c429c72-1ae2-4b7d-9786-57ed9341bd48-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zftfz\" (UID: \"7c429c72-1ae2-4b7d-9786-57ed9341bd48\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zftfz" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616935 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtpcw\" (UniqueName: \"kubernetes.io/projected/4bfeb5de-dcfb-4f0b-b118-112ee3262c18-kube-api-access-qtpcw\") pod \"dns-operator-744455d44c-dvt5z\" (UID: \"4bfeb5de-dcfb-4f0b-b118-112ee3262c18\") " pod="openshift-dns-operator/dns-operator-744455d44c-dvt5z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616962 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/eb6893d2-46c1-46fa-8c40-5bbd900d9fc1-signing-key\") pod \"service-ca-9c57cc56f-jp5fs\" (UID: \"eb6893d2-46c1-46fa-8c40-5bbd900d9fc1\") " pod="openshift-service-ca/service-ca-9c57cc56f-jp5fs" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.616998 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/de7a08c9-b456-475e-a688-56edc28603ac-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hqnns\" (UID: \"de7a08c9-b456-475e-a688-56edc28603ac\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqnns" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617026 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/10918dd9-6412-4172-a0f3-eac3fbd10da1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617050 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/968b0694-9a30-49ca-bee7-d1eadbb31505-config\") pod \"machine-approver-56656f9798-h7srs\" (UID: \"968b0694-9a30-49ca-bee7-d1eadbb31505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h7srs" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617073 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45ba8e80-12bf-472e-a0b7-b25601a9074e-config\") pod \"kube-controller-manager-operator-78b949d7b-887w5\" (UID: \"45ba8e80-12bf-472e-a0b7-b25601a9074e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-887w5" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617099 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-service-ca\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617123 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/10918dd9-6412-4172-a0f3-eac3fbd10da1-audit-policies\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617146 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5484670-b8e2-40b9-a36b-3fb47bb89421-serving-cert\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617171 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/32f920f7-5adb-4cf3-bc1f-3a461b79bb10-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vrvxq\" (UID: \"32f920f7-5adb-4cf3-bc1f-3a461b79bb10\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617194 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9qtj\" (UniqueName: \"kubernetes.io/projected/2fde1815-9a0b-4f85-820c-b482ccdeab3c-kube-api-access-g9qtj\") pod \"ingress-canary-ccd5x\" (UID: \"2fde1815-9a0b-4f85-820c-b482ccdeab3c\") " pod="openshift-ingress-canary/ingress-canary-ccd5x" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617218 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/312c389c-57c0-4709-9320-777e4659d499-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-28f5t\" (UID: \"312c389c-57c0-4709-9320-777e4659d499\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-28f5t" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617242 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617283 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e48b2f0e-a02f-48c6-b513-33537b0108b6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xvd5r\" (UID: \"e48b2f0e-a02f-48c6-b513-33537b0108b6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xvd5r" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617308 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/85928d04-1176-4b1d-8bb7-d55376576225-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2ps29\" (UID: \"85928d04-1176-4b1d-8bb7-d55376576225\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ps29" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617331 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8d3b850f-8917-46af-b713-891a7c979d2e-images\") pod \"machine-api-operator-5694c8668f-8kzxl\" (UID: \"8d3b850f-8917-46af-b713-891a7c979d2e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8kzxl" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617354 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq7dc\" (UniqueName: \"kubernetes.io/projected/091bb385-63c1-443f-9aec-bedd08514ee8-kube-api-access-jq7dc\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617379 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2a1ecdb-4740-43c4-ba73-19d1f15465e1-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rsc75\" (UID: \"f2a1ecdb-4740-43c4-ba73-19d1f15465e1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rsc75" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617404 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66fgc\" (UniqueName: \"kubernetes.io/projected/4aa421fd-173a-4c0d-8c66-ddcb7b173808-kube-api-access-66fgc\") pod \"kube-storage-version-migrator-operator-b67b599dd-jtg8k\" (UID: \"4aa421fd-173a-4c0d-8c66-ddcb7b173808\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jtg8k" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617428 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-console-config\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617450 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md47q\" (UniqueName: \"kubernetes.io/projected/10918dd9-6412-4172-a0f3-eac3fbd10da1-kube-api-access-md47q\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617475 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77sqd\" (UniqueName: \"kubernetes.io/projected/b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a-kube-api-access-77sqd\") pod \"service-ca-operator-777779d784-fxwd8\" (UID: \"b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fxwd8" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617506 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvx6l\" (UniqueName: \"kubernetes.io/projected/85928d04-1176-4b1d-8bb7-d55376576225-kube-api-access-mvx6l\") pod \"cluster-image-registry-operator-dc59b4c8b-2ps29\" (UID: \"85928d04-1176-4b1d-8bb7-d55376576225\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ps29" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617530 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgpx6\" (UniqueName: \"kubernetes.io/projected/5a66393b-1460-4ef7-9c8a-2001a0d187fb-kube-api-access-jgpx6\") pod \"authentication-operator-69f744f599-s8xvt\" (UID: \"5a66393b-1460-4ef7-9c8a-2001a0d187fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8xvt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617554 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eb0cffc3-bf07-411c-822e-aa45c58387cb-service-ca-bundle\") pod \"router-default-5444994796-5gfmz\" (UID: \"eb0cffc3-bf07-411c-822e-aa45c58387cb\") " pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617592 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23111038-ec66-4295-a6eb-b847c52a0c81-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7f58c\" (UID: \"23111038-ec66-4295-a6eb-b847c52a0c81\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7f58c" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617615 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e48b2f0e-a02f-48c6-b513-33537b0108b6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xvd5r\" (UID: \"e48b2f0e-a02f-48c6-b513-33537b0108b6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xvd5r" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617638 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-oauth-serving-cert\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617660 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1d4ef859-1409-451a-a2e1-4c1c8c6d473b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nz99w\" (UID: \"1d4ef859-1409-451a-a2e1-4c1c8c6d473b\") " pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617680 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/751d9838-4c34-4251-b3a6-1d808f98a034-config-volume\") pod \"collect-profiles-29396445-v6dn4\" (UID: \"751d9838-4c34-4251-b3a6-1d808f98a034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617705 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a66393b-1460-4ef7-9c8a-2001a0d187fb-service-ca-bundle\") pod \"authentication-operator-69f744f599-s8xvt\" (UID: \"5a66393b-1460-4ef7-9c8a-2001a0d187fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8xvt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617728 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br4mq\" (UniqueName: \"kubernetes.io/projected/eb6893d2-46c1-46fa-8c40-5bbd900d9fc1-kube-api-access-br4mq\") pod \"service-ca-9c57cc56f-jp5fs\" (UID: \"eb6893d2-46c1-46fa-8c40-5bbd900d9fc1\") " pod="openshift-service-ca/service-ca-9c57cc56f-jp5fs" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617752 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/45ba8e80-12bf-472e-a0b7-b25601a9074e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-887w5\" (UID: \"45ba8e80-12bf-472e-a0b7-b25601a9074e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-887w5" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617774 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617798 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a-serving-cert\") pod \"service-ca-operator-777779d784-fxwd8\" (UID: \"b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fxwd8" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617820 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/40771f13-f771-4e07-8ad9-668a3e48fb3b-etcd-service-ca\") pod \"etcd-operator-b45778765-dhhh4\" (UID: \"40771f13-f771-4e07-8ad9-668a3e48fb3b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617842 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6s8p\" (UniqueName: \"kubernetes.io/projected/7c429c72-1ae2-4b7d-9786-57ed9341bd48-kube-api-access-n6s8p\") pod \"control-plane-machine-set-operator-78cbb6b69f-zftfz\" (UID: \"7c429c72-1ae2-4b7d-9786-57ed9341bd48\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zftfz" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617865 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79g6t\" (UniqueName: \"kubernetes.io/projected/215b4f1f-75ab-4c5c-aca8-26ea0edf5978-kube-api-access-79g6t\") pod \"openshift-config-operator-7777fb866f-rr7gt\" (UID: \"215b4f1f-75ab-4c5c-aca8-26ea0edf5978\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rr7gt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617894 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwpwz\" (UniqueName: \"kubernetes.io/projected/7337e50a-daf8-4f50-9060-ec2a14752444-kube-api-access-kwpwz\") pod \"console-operator-58897d9998-44grv\" (UID: \"7337e50a-daf8-4f50-9060-ec2a14752444\") " pod="openshift-console-operator/console-operator-58897d9998-44grv" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617916 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95th8\" (UniqueName: \"kubernetes.io/projected/61feab0c-5f50-49fe-bf35-de2a5f66bf32-kube-api-access-95th8\") pod \"downloads-7954f5f757-rcvlq\" (UID: \"61feab0c-5f50-49fe-bf35-de2a5f66bf32\") " pod="openshift-console/downloads-7954f5f757-rcvlq" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617939 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a66393b-1460-4ef7-9c8a-2001a0d187fb-config\") pod \"authentication-operator-69f744f599-s8xvt\" (UID: \"5a66393b-1460-4ef7-9c8a-2001a0d187fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8xvt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617962 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea-proxy-tls\") pod \"machine-config-controller-84d6567774-4pxtx\" (UID: \"54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4pxtx" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.617987 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1c05f6d6-e233-46ca-91df-eaba2614018f-metrics-tls\") pod \"ingress-operator-5b745b69d9-92bz6\" (UID: \"1c05f6d6-e233-46ca-91df-eaba2614018f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618011 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j7fh\" (UniqueName: \"kubernetes.io/projected/670a3117-320e-477c-8c4b-a54fe5a4ae9c-kube-api-access-8j7fh\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618035 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/32f920f7-5adb-4cf3-bc1f-3a461b79bb10-proxy-tls\") pod \"machine-config-operator-74547568cd-vrvxq\" (UID: \"32f920f7-5adb-4cf3-bc1f-3a461b79bb10\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618056 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thfxc\" (UniqueName: \"kubernetes.io/projected/3e2ce5bb-1abf-47c6-b6d2-59389118cff0-kube-api-access-thfxc\") pod \"catalog-operator-68c6474976-btgsr\" (UID: \"3e2ce5bb-1abf-47c6-b6d2-59389118cff0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618080 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a66393b-1460-4ef7-9c8a-2001a0d187fb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-s8xvt\" (UID: \"5a66393b-1460-4ef7-9c8a-2001a0d187fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8xvt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618103 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618126 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618150 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ns7n\" (UniqueName: \"kubernetes.io/projected/7ab520a0-54e7-4923-867c-215fffacf270-kube-api-access-2ns7n\") pod \"multus-admission-controller-857f4d67dd-7bmhw\" (UID: \"7ab520a0-54e7-4923-867c-215fffacf270\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7bmhw" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618172 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/10918dd9-6412-4172-a0f3-eac3fbd10da1-etcd-client\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618211 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618237 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe-profile-collector-cert\") pod \"olm-operator-6b444d44fb-jlvsb\" (UID: \"5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618285 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23111038-ec66-4295-a6eb-b847c52a0c81-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7f58c\" (UID: \"23111038-ec66-4295-a6eb-b847c52a0c81\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7f58c" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618322 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/10918dd9-6412-4172-a0f3-eac3fbd10da1-audit-dir\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618359 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/091bb385-63c1-443f-9aec-bedd08514ee8-serving-cert\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618382 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/353ed841-0d3b-46df-beec-1906f6d801fd-client-ca\") pod \"route-controller-manager-6576b87f9c-4lz4x\" (UID: \"353ed841-0d3b-46df-beec-1906f6d801fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618403 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c9472804-f8b1-440a-b12e-d351630ed795-tmpfs\") pod \"packageserver-d55dfcdfc-k27q8\" (UID: \"c9472804-f8b1-440a-b12e-d351630ed795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618430 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/62bdffb2-a1c7-4960-81aa-b570834f78cd-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-j7ttg\" (UID: \"62bdffb2-a1c7-4960-81aa-b570834f78cd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j7ttg" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618454 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhqfr\" (UniqueName: \"kubernetes.io/projected/40771f13-f771-4e07-8ad9-668a3e48fb3b-kube-api-access-qhqfr\") pod \"etcd-operator-b45778765-dhhh4\" (UID: \"40771f13-f771-4e07-8ad9-668a3e48fb3b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618474 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a33da900-6b7c-4157-8a6e-fde1cddf51ca-mountpoint-dir\") pod \"csi-hostpathplugin-8zp8h\" (UID: \"a33da900-6b7c-4157-8a6e-fde1cddf51ca\") " pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618500 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/10918dd9-6412-4172-a0f3-eac3fbd10da1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618524 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/10918dd9-6412-4172-a0f3-eac3fbd10da1-encryption-config\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618545 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/eb6893d2-46c1-46fa-8c40-5bbd900d9fc1-signing-cabundle\") pod \"service-ca-9c57cc56f-jp5fs\" (UID: \"eb6893d2-46c1-46fa-8c40-5bbd900d9fc1\") " pod="openshift-service-ca/service-ca-9c57cc56f-jp5fs" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618570 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7337e50a-daf8-4f50-9060-ec2a14752444-serving-cert\") pod \"console-operator-58897d9998-44grv\" (UID: \"7337e50a-daf8-4f50-9060-ec2a14752444\") " pod="openshift-console-operator/console-operator-58897d9998-44grv" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618585 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/091bb385-63c1-443f-9aec-bedd08514ee8-node-pullsecrets\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618593 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a33da900-6b7c-4157-8a6e-fde1cddf51ca-registration-dir\") pod \"csi-hostpathplugin-8zp8h\" (UID: \"a33da900-6b7c-4157-8a6e-fde1cddf51ca\") " pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618640 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a33da900-6b7c-4157-8a6e-fde1cddf51ca-plugins-dir\") pod \"csi-hostpathplugin-8zp8h\" (UID: \"a33da900-6b7c-4157-8a6e-fde1cddf51ca\") " pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618667 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d3b850f-8917-46af-b713-891a7c979d2e-config\") pod \"machine-api-operator-5694c8668f-8kzxl\" (UID: \"8d3b850f-8917-46af-b713-891a7c979d2e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8kzxl" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618691 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c9472804-f8b1-440a-b12e-d351630ed795-apiservice-cert\") pod \"packageserver-d55dfcdfc-k27q8\" (UID: \"c9472804-f8b1-440a-b12e-d351630ed795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618694 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/091bb385-63c1-443f-9aec-bedd08514ee8-audit-dir\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618718 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/670a3117-320e-477c-8c4b-a54fe5a4ae9c-console-serving-cert\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618742 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7337e50a-daf8-4f50-9060-ec2a14752444-config\") pod \"console-operator-58897d9998-44grv\" (UID: \"7337e50a-daf8-4f50-9060-ec2a14752444\") " pod="openshift-console-operator/console-operator-58897d9998-44grv" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618767 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6f239350-c7bc-4fe4-9c24-ca0070a0775c-certs\") pod \"machine-config-server-z7mtd\" (UID: \"6f239350-c7bc-4fe4-9c24-ca0070a0775c\") " pod="openshift-machine-config-operator/machine-config-server-z7mtd" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618791 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618819 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1c05f6d6-e233-46ca-91df-eaba2614018f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-92bz6\" (UID: \"1c05f6d6-e233-46ca-91df-eaba2614018f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618844 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vlnj\" (UniqueName: \"kubernetes.io/projected/32f920f7-5adb-4cf3-bc1f-3a461b79bb10-kube-api-access-7vlnj\") pod \"machine-config-operator-74547568cd-vrvxq\" (UID: \"32f920f7-5adb-4cf3-bc1f-3a461b79bb10\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618866 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9r5z\" (UniqueName: \"kubernetes.io/projected/6f239350-c7bc-4fe4-9c24-ca0070a0775c-kube-api-access-x9r5z\") pod \"machine-config-server-z7mtd\" (UID: \"6f239350-c7bc-4fe4-9c24-ca0070a0775c\") " pod="openshift-machine-config-operator/machine-config-server-z7mtd" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618891 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0dcaa80a-8993-4452-9acb-b8010eade7c6-audit-dir\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618914 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/968b0694-9a30-49ca-bee7-d1eadbb31505-machine-approver-tls\") pod \"machine-approver-56656f9798-h7srs\" (UID: \"968b0694-9a30-49ca-bee7-d1eadbb31505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h7srs" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618938 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1d4ef859-1409-451a-a2e1-4c1c8c6d473b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nz99w\" (UID: \"1d4ef859-1409-451a-a2e1-4c1c8c6d473b\") " pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618960 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3e2ce5bb-1abf-47c6-b6d2-59389118cff0-srv-cert\") pod \"catalog-operator-68c6474976-btgsr\" (UID: \"3e2ce5bb-1abf-47c6-b6d2-59389118cff0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.618982 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-config\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.619006 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/670a3117-320e-477c-8c4b-a54fe5a4ae9c-console-oauth-config\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.619026 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/091bb385-63c1-443f-9aec-bedd08514ee8-config\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.619094 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4aa421fd-173a-4c0d-8c66-ddcb7b173808-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jtg8k\" (UID: \"4aa421fd-173a-4c0d-8c66-ddcb7b173808\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jtg8k" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.619121 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2fde1815-9a0b-4f85-820c-b482ccdeab3c-cert\") pod \"ingress-canary-ccd5x\" (UID: \"2fde1815-9a0b-4f85-820c-b482ccdeab3c\") " pod="openshift-ingress-canary/ingress-canary-ccd5x" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.619200 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcll2\" (UniqueName: \"kubernetes.io/projected/8d3b850f-8917-46af-b713-891a7c979d2e-kube-api-access-tcll2\") pod \"machine-api-operator-5694c8668f-8kzxl\" (UID: \"8d3b850f-8917-46af-b713-891a7c979d2e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8kzxl" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.619268 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fpr4\" (UniqueName: \"kubernetes.io/projected/eb0cffc3-bf07-411c-822e-aa45c58387cb-kube-api-access-8fpr4\") pod \"router-default-5444994796-5gfmz\" (UID: \"eb0cffc3-bf07-411c-822e-aa45c58387cb\") " pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.619406 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/968b0694-9a30-49ca-bee7-d1eadbb31505-auth-proxy-config\") pod \"machine-approver-56656f9798-h7srs\" (UID: \"968b0694-9a30-49ca-bee7-d1eadbb31505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h7srs" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620217 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10918dd9-6412-4172-a0f3-eac3fbd10da1-serving-cert\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620277 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40771f13-f771-4e07-8ad9-668a3e48fb3b-serving-cert\") pod \"etcd-operator-b45778765-dhhh4\" (UID: \"40771f13-f771-4e07-8ad9-668a3e48fb3b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620331 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1c05f6d6-e233-46ca-91df-eaba2614018f-trusted-ca\") pod \"ingress-operator-5b745b69d9-92bz6\" (UID: \"1c05f6d6-e233-46ca-91df-eaba2614018f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620358 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4bfeb5de-dcfb-4f0b-b118-112ee3262c18-metrics-tls\") pod \"dns-operator-744455d44c-dvt5z\" (UID: \"4bfeb5de-dcfb-4f0b-b118-112ee3262c18\") " pod="openshift-dns-operator/dns-operator-744455d44c-dvt5z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620381 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb0cffc3-bf07-411c-822e-aa45c58387cb-metrics-certs\") pod \"router-default-5444994796-5gfmz\" (UID: \"eb0cffc3-bf07-411c-822e-aa45c58387cb\") " pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620407 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/215b4f1f-75ab-4c5c-aca8-26ea0edf5978-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rr7gt\" (UID: \"215b4f1f-75ab-4c5c-aca8-26ea0edf5978\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rr7gt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620417 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0dcaa80a-8993-4452-9acb-b8010eade7c6-audit-dir\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620439 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/312c389c-57c0-4709-9320-777e4659d499-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-28f5t\" (UID: \"312c389c-57c0-4709-9320-777e4659d499\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-28f5t" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620484 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe-srv-cert\") pod \"olm-operator-6b444d44fb-jlvsb\" (UID: \"5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620516 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gsdk\" (UniqueName: \"kubernetes.io/projected/1c05f6d6-e233-46ca-91df-eaba2614018f-kube-api-access-8gsdk\") pod \"ingress-operator-5b745b69d9-92bz6\" (UID: \"1c05f6d6-e233-46ca-91df-eaba2614018f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620527 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45ba8e80-12bf-472e-a0b7-b25601a9074e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-887w5\" (UID: \"45ba8e80-12bf-472e-a0b7-b25601a9074e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-887w5" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620543 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v28zk\" (UniqueName: \"kubernetes.io/projected/c9472804-f8b1-440a-b12e-d351630ed795-kube-api-access-v28zk\") pod \"packageserver-d55dfcdfc-k27q8\" (UID: \"c9472804-f8b1-440a-b12e-d351630ed795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620571 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdkln\" (UniqueName: \"kubernetes.io/projected/968b0694-9a30-49ca-bee7-d1eadbb31505-kube-api-access-jdkln\") pod \"machine-approver-56656f9798-h7srs\" (UID: \"968b0694-9a30-49ca-bee7-d1eadbb31505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h7srs" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620596 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620620 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7ab520a0-54e7-4923-867c-215fffacf270-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7bmhw\" (UID: \"7ab520a0-54e7-4923-867c-215fffacf270\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7bmhw" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620649 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/215b4f1f-75ab-4c5c-aca8-26ea0edf5978-serving-cert\") pod \"openshift-config-operator-7777fb866f-rr7gt\" (UID: \"215b4f1f-75ab-4c5c-aca8-26ea0edf5978\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rr7gt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620672 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a33da900-6b7c-4157-8a6e-fde1cddf51ca-csi-data-dir\") pod \"csi-hostpathplugin-8zp8h\" (UID: \"a33da900-6b7c-4157-8a6e-fde1cddf51ca\") " pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620701 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/091bb385-63c1-443f-9aec-bedd08514ee8-image-import-ca\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620725 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4aa421fd-173a-4c0d-8c66-ddcb7b173808-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jtg8k\" (UID: \"4aa421fd-173a-4c0d-8c66-ddcb7b173808\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jtg8k" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620747 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c9472804-f8b1-440a-b12e-d351630ed795-webhook-cert\") pod \"packageserver-d55dfcdfc-k27q8\" (UID: \"c9472804-f8b1-440a-b12e-d351630ed795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620774 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtddb\" (UniqueName: \"kubernetes.io/projected/54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea-kube-api-access-wtddb\") pod \"machine-config-controller-84d6567774-4pxtx\" (UID: \"54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4pxtx" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620791 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/091bb385-63c1-443f-9aec-bedd08514ee8-encryption-config\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620800 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6f239350-c7bc-4fe4-9c24-ca0070a0775c-node-bootstrap-token\") pod \"machine-config-server-z7mtd\" (UID: \"6f239350-c7bc-4fe4-9c24-ca0070a0775c\") " pod="openshift-machine-config-operator/machine-config-server-z7mtd" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620830 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/312c389c-57c0-4709-9320-777e4659d499-config\") pod \"kube-apiserver-operator-766d6c64bb-28f5t\" (UID: \"312c389c-57c0-4709-9320-777e4659d499\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-28f5t" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620854 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8vfr\" (UniqueName: \"kubernetes.io/projected/5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe-kube-api-access-j8vfr\") pod \"olm-operator-6b444d44fb-jlvsb\" (UID: \"5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620880 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620907 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23111038-ec66-4295-a6eb-b847c52a0c81-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7f58c\" (UID: \"23111038-ec66-4295-a6eb-b847c52a0c81\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7f58c" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620932 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgrvf\" (UniqueName: \"kubernetes.io/projected/de7a08c9-b456-475e-a688-56edc28603ac-kube-api-access-hgrvf\") pod \"package-server-manager-789f6589d5-hqnns\" (UID: \"de7a08c9-b456-475e-a688-56edc28603ac\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqnns" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620956 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/32f920f7-5adb-4cf3-bc1f-3a461b79bb10-images\") pod \"machine-config-operator-74547568cd-vrvxq\" (UID: \"32f920f7-5adb-4cf3-bc1f-3a461b79bb10\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.620983 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/85928d04-1176-4b1d-8bb7-d55376576225-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2ps29\" (UID: \"85928d04-1176-4b1d-8bb7-d55376576225\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ps29" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.621011 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a66393b-1460-4ef7-9c8a-2001a0d187fb-serving-cert\") pod \"authentication-operator-69f744f599-s8xvt\" (UID: \"5a66393b-1460-4ef7-9c8a-2001a0d187fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8xvt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.621091 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a66393b-1460-4ef7-9c8a-2001a0d187fb-config\") pod \"authentication-operator-69f744f599-s8xvt\" (UID: \"5a66393b-1460-4ef7-9c8a-2001a0d187fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8xvt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.621214 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8d3b850f-8917-46af-b713-891a7c979d2e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-8kzxl\" (UID: \"8d3b850f-8917-46af-b713-891a7c979d2e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8kzxl" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.621540 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a66393b-1460-4ef7-9c8a-2001a0d187fb-service-ca-bundle\") pod \"authentication-operator-69f744f599-s8xvt\" (UID: \"5a66393b-1460-4ef7-9c8a-2001a0d187fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8xvt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.621633 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.621868 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7337e50a-daf8-4f50-9060-ec2a14752444-config\") pod \"console-operator-58897d9998-44grv\" (UID: \"7337e50a-daf8-4f50-9060-ec2a14752444\") " pod="openshift-console-operator/console-operator-58897d9998-44grv" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.622977 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/968b0694-9a30-49ca-bee7-d1eadbb31505-config\") pod \"machine-approver-56656f9798-h7srs\" (UID: \"968b0694-9a30-49ca-bee7-d1eadbb31505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h7srs" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.623137 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/10918dd9-6412-4172-a0f3-eac3fbd10da1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.623195 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7337e50a-daf8-4f50-9060-ec2a14752444-trusted-ca\") pod \"console-operator-58897d9998-44grv\" (UID: \"7337e50a-daf8-4f50-9060-ec2a14752444\") " pod="openshift-console-operator/console-operator-58897d9998-44grv" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.623506 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e48b2f0e-a02f-48c6-b513-33537b0108b6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xvd5r\" (UID: \"e48b2f0e-a02f-48c6-b513-33537b0108b6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xvd5r" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.624084 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/85928d04-1176-4b1d-8bb7-d55376576225-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2ps29\" (UID: \"85928d04-1176-4b1d-8bb7-d55376576225\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ps29" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.624676 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1c05f6d6-e233-46ca-91df-eaba2614018f-trusted-ca\") pod \"ingress-operator-5b745b69d9-92bz6\" (UID: \"1c05f6d6-e233-46ca-91df-eaba2614018f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.624973 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a66393b-1460-4ef7-9c8a-2001a0d187fb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-s8xvt\" (UID: \"5a66393b-1460-4ef7-9c8a-2001a0d187fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8xvt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.625280 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.625554 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23111038-ec66-4295-a6eb-b847c52a0c81-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7f58c\" (UID: \"23111038-ec66-4295-a6eb-b847c52a0c81\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7f58c" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.625940 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/10918dd9-6412-4172-a0f3-eac3fbd10da1-etcd-client\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.626145 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8d3b850f-8917-46af-b713-891a7c979d2e-images\") pod \"machine-api-operator-5694c8668f-8kzxl\" (UID: \"8d3b850f-8917-46af-b713-891a7c979d2e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8kzxl" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.626231 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/312c389c-57c0-4709-9320-777e4659d499-config\") pod \"kube-apiserver-operator-766d6c64bb-28f5t\" (UID: \"312c389c-57c0-4709-9320-777e4659d499\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-28f5t" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.626556 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/312c389c-57c0-4709-9320-777e4659d499-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-28f5t\" (UID: \"312c389c-57c0-4709-9320-777e4659d499\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-28f5t" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.627070 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/670a3117-320e-477c-8c4b-a54fe5a4ae9c-console-serving-cert\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.627123 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4bfeb5de-dcfb-4f0b-b118-112ee3262c18-metrics-tls\") pod \"dns-operator-744455d44c-dvt5z\" (UID: \"4bfeb5de-dcfb-4f0b-b118-112ee3262c18\") " pod="openshift-dns-operator/dns-operator-744455d44c-dvt5z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.627386 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.627489 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23111038-ec66-4295-a6eb-b847c52a0c81-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7f58c\" (UID: \"23111038-ec66-4295-a6eb-b847c52a0c81\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7f58c" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.627523 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/091bb385-63c1-443f-9aec-bedd08514ee8-image-import-ca\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.627425 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/353ed841-0d3b-46df-beec-1906f6d801fd-config\") pod \"route-controller-manager-6576b87f9c-4lz4x\" (UID: \"353ed841-0d3b-46df-beec-1906f6d801fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.627593 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/10918dd9-6412-4172-a0f3-eac3fbd10da1-audit-dir\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.628081 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/10918dd9-6412-4172-a0f3-eac3fbd10da1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.628244 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e48b2f0e-a02f-48c6-b513-33537b0108b6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xvd5r\" (UID: \"e48b2f0e-a02f-48c6-b513-33537b0108b6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xvd5r" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.628811 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.628895 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/091bb385-63c1-443f-9aec-bedd08514ee8-audit\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.629050 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/10918dd9-6412-4172-a0f3-eac3fbd10da1-audit-policies\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.629109 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/353ed841-0d3b-46df-beec-1906f6d801fd-client-ca\") pod \"route-controller-manager-6576b87f9c-4lz4x\" (UID: \"353ed841-0d3b-46df-beec-1906f6d801fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.629279 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45ba8e80-12bf-472e-a0b7-b25601a9074e-config\") pod \"kube-controller-manager-operator-78b949d7b-887w5\" (UID: \"45ba8e80-12bf-472e-a0b7-b25601a9074e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-887w5" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.629287 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.629491 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-audit-policies\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.629504 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/091bb385-63c1-443f-9aec-bedd08514ee8-etcd-client\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.629527 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1c05f6d6-e233-46ca-91df-eaba2614018f-metrics-tls\") pod \"ingress-operator-5b745b69d9-92bz6\" (UID: \"1c05f6d6-e233-46ca-91df-eaba2614018f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.629792 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.629934 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/091bb385-63c1-443f-9aec-bedd08514ee8-etcd-serving-ca\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.630065 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.630432 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/670a3117-320e-477c-8c4b-a54fe5a4ae9c-console-oauth-config\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.630549 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.630672 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10918dd9-6412-4172-a0f3-eac3fbd10da1-serving-cert\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.631271 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a66393b-1460-4ef7-9c8a-2001a0d187fb-serving-cert\") pod \"authentication-operator-69f744f599-s8xvt\" (UID: \"5a66393b-1460-4ef7-9c8a-2001a0d187fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8xvt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.631571 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/10918dd9-6412-4172-a0f3-eac3fbd10da1-encryption-config\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.631990 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/091bb385-63c1-443f-9aec-bedd08514ee8-config\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.632169 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/968b0694-9a30-49ca-bee7-d1eadbb31505-machine-approver-tls\") pod \"machine-approver-56656f9798-h7srs\" (UID: \"968b0694-9a30-49ca-bee7-d1eadbb31505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h7srs" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.633571 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7337e50a-daf8-4f50-9060-ec2a14752444-serving-cert\") pod \"console-operator-58897d9998-44grv\" (UID: \"7337e50a-daf8-4f50-9060-ec2a14752444\") " pod="openshift-console-operator/console-operator-58897d9998-44grv" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.633781 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.633991 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.634136 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/091bb385-63c1-443f-9aec-bedd08514ee8-serving-cert\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.634659 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.634815 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/62bdffb2-a1c7-4960-81aa-b570834f78cd-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-j7ttg\" (UID: \"62bdffb2-a1c7-4960-81aa-b570834f78cd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j7ttg" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.634823 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/85928d04-1176-4b1d-8bb7-d55376576225-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2ps29\" (UID: \"85928d04-1176-4b1d-8bb7-d55376576225\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ps29" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.635399 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.636099 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.637330 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-service-ca\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.637713 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/353ed841-0d3b-46df-beec-1906f6d801fd-serving-cert\") pod \"route-controller-manager-6576b87f9c-4lz4x\" (UID: \"353ed841-0d3b-46df-beec-1906f6d801fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.644290 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.651042 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-console-config\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.675398 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.681588 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-oauth-serving-cert\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.694905 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.699767 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-trusted-ca-bundle\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.705237 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.721534 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/eb0cffc3-bf07-411c-822e-aa45c58387cb-default-certificate\") pod \"router-default-5444994796-5gfmz\" (UID: \"eb0cffc3-bf07-411c-822e-aa45c58387cb\") " pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.721699 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk6gh\" (UniqueName: \"kubernetes.io/projected/a33da900-6b7c-4157-8a6e-fde1cddf51ca-kube-api-access-zk6gh\") pod \"csi-hostpathplugin-8zp8h\" (UID: \"a33da900-6b7c-4157-8a6e-fde1cddf51ca\") " pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.721819 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/40771f13-f771-4e07-8ad9-668a3e48fb3b-etcd-client\") pod \"etcd-operator-b45778765-dhhh4\" (UID: \"40771f13-f771-4e07-8ad9-668a3e48fb3b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.721925 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpq48\" (UniqueName: \"kubernetes.io/projected/1d4ef859-1409-451a-a2e1-4c1c8c6d473b-kube-api-access-kpq48\") pod \"marketplace-operator-79b997595-nz99w\" (UID: \"1d4ef859-1409-451a-a2e1-4c1c8c6d473b\") " pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.722053 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a33da900-6b7c-4157-8a6e-fde1cddf51ca-socket-dir\") pod \"csi-hostpathplugin-8zp8h\" (UID: \"a33da900-6b7c-4157-8a6e-fde1cddf51ca\") " pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.722173 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2a1ecdb-4740-43c4-ba73-19d1f15465e1-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rsc75\" (UID: \"f2a1ecdb-4740-43c4-ba73-19d1f15465e1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rsc75" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.722311 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/751d9838-4c34-4251-b3a6-1d808f98a034-secret-volume\") pod \"collect-profiles-29396445-v6dn4\" (UID: \"751d9838-4c34-4251-b3a6-1d808f98a034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.722435 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a-config\") pod \"service-ca-operator-777779d784-fxwd8\" (UID: \"b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fxwd8" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.722546 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40771f13-f771-4e07-8ad9-668a3e48fb3b-config\") pod \"etcd-operator-b45778765-dhhh4\" (UID: \"40771f13-f771-4e07-8ad9-668a3e48fb3b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.722466 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a33da900-6b7c-4157-8a6e-fde1cddf51ca-socket-dir\") pod \"csi-hostpathplugin-8zp8h\" (UID: \"a33da900-6b7c-4157-8a6e-fde1cddf51ca\") " pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.722651 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cccc9\" (UniqueName: \"kubernetes.io/projected/f2a1ecdb-4740-43c4-ba73-19d1f15465e1-kube-api-access-cccc9\") pod \"openshift-controller-manager-operator-756b6f6bc6-rsc75\" (UID: \"f2a1ecdb-4740-43c4-ba73-19d1f15465e1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rsc75" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.723051 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/40771f13-f771-4e07-8ad9-668a3e48fb3b-etcd-ca\") pod \"etcd-operator-b45778765-dhhh4\" (UID: \"40771f13-f771-4e07-8ad9-668a3e48fb3b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.723095 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-4pxtx\" (UID: \"54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4pxtx" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.724013 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-4pxtx\" (UID: \"54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4pxtx" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.724069 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/eb0cffc3-bf07-411c-822e-aa45c58387cb-stats-auth\") pod \"router-default-5444994796-5gfmz\" (UID: \"eb0cffc3-bf07-411c-822e-aa45c58387cb\") " pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.724119 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7c429c72-1ae2-4b7d-9786-57ed9341bd48-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zftfz\" (UID: \"7c429c72-1ae2-4b7d-9786-57ed9341bd48\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zftfz" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.724160 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/eb6893d2-46c1-46fa-8c40-5bbd900d9fc1-signing-key\") pod \"service-ca-9c57cc56f-jp5fs\" (UID: \"eb6893d2-46c1-46fa-8c40-5bbd900d9fc1\") " pod="openshift-service-ca/service-ca-9c57cc56f-jp5fs" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.724184 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/de7a08c9-b456-475e-a688-56edc28603ac-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hqnns\" (UID: \"de7a08c9-b456-475e-a688-56edc28603ac\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqnns" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.724219 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/32f920f7-5adb-4cf3-bc1f-3a461b79bb10-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vrvxq\" (UID: \"32f920f7-5adb-4cf3-bc1f-3a461b79bb10\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.724242 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9qtj\" (UniqueName: \"kubernetes.io/projected/2fde1815-9a0b-4f85-820c-b482ccdeab3c-kube-api-access-g9qtj\") pod \"ingress-canary-ccd5x\" (UID: \"2fde1815-9a0b-4f85-820c-b482ccdeab3c\") " pod="openshift-ingress-canary/ingress-canary-ccd5x" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.724296 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66fgc\" (UniqueName: \"kubernetes.io/projected/4aa421fd-173a-4c0d-8c66-ddcb7b173808-kube-api-access-66fgc\") pod \"kube-storage-version-migrator-operator-b67b599dd-jtg8k\" (UID: \"4aa421fd-173a-4c0d-8c66-ddcb7b173808\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jtg8k" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.724333 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2a1ecdb-4740-43c4-ba73-19d1f15465e1-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rsc75\" (UID: \"f2a1ecdb-4740-43c4-ba73-19d1f15465e1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rsc75" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.724365 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77sqd\" (UniqueName: \"kubernetes.io/projected/b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a-kube-api-access-77sqd\") pod \"service-ca-operator-777779d784-fxwd8\" (UID: \"b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fxwd8" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.724404 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eb0cffc3-bf07-411c-822e-aa45c58387cb-service-ca-bundle\") pod \"router-default-5444994796-5gfmz\" (UID: \"eb0cffc3-bf07-411c-822e-aa45c58387cb\") " pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.724786 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.725187 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/32f920f7-5adb-4cf3-bc1f-3a461b79bb10-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vrvxq\" (UID: \"32f920f7-5adb-4cf3-bc1f-3a461b79bb10\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.725329 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1d4ef859-1409-451a-a2e1-4c1c8c6d473b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nz99w\" (UID: \"1d4ef859-1409-451a-a2e1-4c1c8c6d473b\") " pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.725361 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/751d9838-4c34-4251-b3a6-1d808f98a034-config-volume\") pod \"collect-profiles-29396445-v6dn4\" (UID: \"751d9838-4c34-4251-b3a6-1d808f98a034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.725485 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br4mq\" (UniqueName: \"kubernetes.io/projected/eb6893d2-46c1-46fa-8c40-5bbd900d9fc1-kube-api-access-br4mq\") pod \"service-ca-9c57cc56f-jp5fs\" (UID: \"eb6893d2-46c1-46fa-8c40-5bbd900d9fc1\") " pod="openshift-service-ca/service-ca-9c57cc56f-jp5fs" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.725566 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a-serving-cert\") pod \"service-ca-operator-777779d784-fxwd8\" (UID: \"b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fxwd8" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.725687 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/40771f13-f771-4e07-8ad9-668a3e48fb3b-etcd-service-ca\") pod \"etcd-operator-b45778765-dhhh4\" (UID: \"40771f13-f771-4e07-8ad9-668a3e48fb3b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.725795 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6s8p\" (UniqueName: \"kubernetes.io/projected/7c429c72-1ae2-4b7d-9786-57ed9341bd48-kube-api-access-n6s8p\") pod \"control-plane-machine-set-operator-78cbb6b69f-zftfz\" (UID: \"7c429c72-1ae2-4b7d-9786-57ed9341bd48\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zftfz" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.725850 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79g6t\" (UniqueName: \"kubernetes.io/projected/215b4f1f-75ab-4c5c-aca8-26ea0edf5978-kube-api-access-79g6t\") pod \"openshift-config-operator-7777fb866f-rr7gt\" (UID: \"215b4f1f-75ab-4c5c-aca8-26ea0edf5978\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rr7gt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.725916 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea-proxy-tls\") pod \"machine-config-controller-84d6567774-4pxtx\" (UID: \"54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4pxtx" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.726076 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/32f920f7-5adb-4cf3-bc1f-3a461b79bb10-proxy-tls\") pod \"machine-config-operator-74547568cd-vrvxq\" (UID: \"32f920f7-5adb-4cf3-bc1f-3a461b79bb10\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.726180 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thfxc\" (UniqueName: \"kubernetes.io/projected/3e2ce5bb-1abf-47c6-b6d2-59389118cff0-kube-api-access-thfxc\") pod \"catalog-operator-68c6474976-btgsr\" (UID: \"3e2ce5bb-1abf-47c6-b6d2-59389118cff0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.726236 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ns7n\" (UniqueName: \"kubernetes.io/projected/7ab520a0-54e7-4923-867c-215fffacf270-kube-api-access-2ns7n\") pod \"multus-admission-controller-857f4d67dd-7bmhw\" (UID: \"7ab520a0-54e7-4923-867c-215fffacf270\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7bmhw" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.726411 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe-profile-collector-cert\") pod \"olm-operator-6b444d44fb-jlvsb\" (UID: \"5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.726526 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c9472804-f8b1-440a-b12e-d351630ed795-tmpfs\") pod \"packageserver-d55dfcdfc-k27q8\" (UID: \"c9472804-f8b1-440a-b12e-d351630ed795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.727303 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c9472804-f8b1-440a-b12e-d351630ed795-tmpfs\") pod \"packageserver-d55dfcdfc-k27q8\" (UID: \"c9472804-f8b1-440a-b12e-d351630ed795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729088 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhqfr\" (UniqueName: \"kubernetes.io/projected/40771f13-f771-4e07-8ad9-668a3e48fb3b-kube-api-access-qhqfr\") pod \"etcd-operator-b45778765-dhhh4\" (UID: \"40771f13-f771-4e07-8ad9-668a3e48fb3b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729137 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a33da900-6b7c-4157-8a6e-fde1cddf51ca-mountpoint-dir\") pod \"csi-hostpathplugin-8zp8h\" (UID: \"a33da900-6b7c-4157-8a6e-fde1cddf51ca\") " pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729165 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/eb6893d2-46c1-46fa-8c40-5bbd900d9fc1-signing-cabundle\") pod \"service-ca-9c57cc56f-jp5fs\" (UID: \"eb6893d2-46c1-46fa-8c40-5bbd900d9fc1\") " pod="openshift-service-ca/service-ca-9c57cc56f-jp5fs" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729190 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c9472804-f8b1-440a-b12e-d351630ed795-apiservice-cert\") pod \"packageserver-d55dfcdfc-k27q8\" (UID: \"c9472804-f8b1-440a-b12e-d351630ed795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729212 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a33da900-6b7c-4157-8a6e-fde1cddf51ca-registration-dir\") pod \"csi-hostpathplugin-8zp8h\" (UID: \"a33da900-6b7c-4157-8a6e-fde1cddf51ca\") " pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729234 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a33da900-6b7c-4157-8a6e-fde1cddf51ca-plugins-dir\") pod \"csi-hostpathplugin-8zp8h\" (UID: \"a33da900-6b7c-4157-8a6e-fde1cddf51ca\") " pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729291 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6f239350-c7bc-4fe4-9c24-ca0070a0775c-certs\") pod \"machine-config-server-z7mtd\" (UID: \"6f239350-c7bc-4fe4-9c24-ca0070a0775c\") " pod="openshift-machine-config-operator/machine-config-server-z7mtd" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729328 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vlnj\" (UniqueName: \"kubernetes.io/projected/32f920f7-5adb-4cf3-bc1f-3a461b79bb10-kube-api-access-7vlnj\") pod \"machine-config-operator-74547568cd-vrvxq\" (UID: \"32f920f7-5adb-4cf3-bc1f-3a461b79bb10\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729353 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9r5z\" (UniqueName: \"kubernetes.io/projected/6f239350-c7bc-4fe4-9c24-ca0070a0775c-kube-api-access-x9r5z\") pod \"machine-config-server-z7mtd\" (UID: \"6f239350-c7bc-4fe4-9c24-ca0070a0775c\") " pod="openshift-machine-config-operator/machine-config-server-z7mtd" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729382 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1d4ef859-1409-451a-a2e1-4c1c8c6d473b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nz99w\" (UID: \"1d4ef859-1409-451a-a2e1-4c1c8c6d473b\") " pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729402 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3e2ce5bb-1abf-47c6-b6d2-59389118cff0-srv-cert\") pod \"catalog-operator-68c6474976-btgsr\" (UID: \"3e2ce5bb-1abf-47c6-b6d2-59389118cff0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729433 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4aa421fd-173a-4c0d-8c66-ddcb7b173808-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jtg8k\" (UID: \"4aa421fd-173a-4c0d-8c66-ddcb7b173808\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jtg8k" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729457 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fpr4\" (UniqueName: \"kubernetes.io/projected/eb0cffc3-bf07-411c-822e-aa45c58387cb-kube-api-access-8fpr4\") pod \"router-default-5444994796-5gfmz\" (UID: \"eb0cffc3-bf07-411c-822e-aa45c58387cb\") " pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729480 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2fde1815-9a0b-4f85-820c-b482ccdeab3c-cert\") pod \"ingress-canary-ccd5x\" (UID: \"2fde1815-9a0b-4f85-820c-b482ccdeab3c\") " pod="openshift-ingress-canary/ingress-canary-ccd5x" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729523 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40771f13-f771-4e07-8ad9-668a3e48fb3b-serving-cert\") pod \"etcd-operator-b45778765-dhhh4\" (UID: \"40771f13-f771-4e07-8ad9-668a3e48fb3b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729552 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb0cffc3-bf07-411c-822e-aa45c58387cb-metrics-certs\") pod \"router-default-5444994796-5gfmz\" (UID: \"eb0cffc3-bf07-411c-822e-aa45c58387cb\") " pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729576 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/215b4f1f-75ab-4c5c-aca8-26ea0edf5978-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rr7gt\" (UID: \"215b4f1f-75ab-4c5c-aca8-26ea0edf5978\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rr7gt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729614 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe-srv-cert\") pod \"olm-operator-6b444d44fb-jlvsb\" (UID: \"5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729640 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v28zk\" (UniqueName: \"kubernetes.io/projected/c9472804-f8b1-440a-b12e-d351630ed795-kube-api-access-v28zk\") pod \"packageserver-d55dfcdfc-k27q8\" (UID: \"c9472804-f8b1-440a-b12e-d351630ed795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729677 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7ab520a0-54e7-4923-867c-215fffacf270-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7bmhw\" (UID: \"7ab520a0-54e7-4923-867c-215fffacf270\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7bmhw" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729698 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/215b4f1f-75ab-4c5c-aca8-26ea0edf5978-serving-cert\") pod \"openshift-config-operator-7777fb866f-rr7gt\" (UID: \"215b4f1f-75ab-4c5c-aca8-26ea0edf5978\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rr7gt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729723 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a33da900-6b7c-4157-8a6e-fde1cddf51ca-csi-data-dir\") pod \"csi-hostpathplugin-8zp8h\" (UID: \"a33da900-6b7c-4157-8a6e-fde1cddf51ca\") " pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729748 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4aa421fd-173a-4c0d-8c66-ddcb7b173808-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jtg8k\" (UID: \"4aa421fd-173a-4c0d-8c66-ddcb7b173808\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jtg8k" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729772 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c9472804-f8b1-440a-b12e-d351630ed795-webhook-cert\") pod \"packageserver-d55dfcdfc-k27q8\" (UID: \"c9472804-f8b1-440a-b12e-d351630ed795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729798 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8vfr\" (UniqueName: \"kubernetes.io/projected/5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe-kube-api-access-j8vfr\") pod \"olm-operator-6b444d44fb-jlvsb\" (UID: \"5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729821 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtddb\" (UniqueName: \"kubernetes.io/projected/54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea-kube-api-access-wtddb\") pod \"machine-config-controller-84d6567774-4pxtx\" (UID: \"54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4pxtx" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729845 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6f239350-c7bc-4fe4-9c24-ca0070a0775c-node-bootstrap-token\") pod \"machine-config-server-z7mtd\" (UID: \"6f239350-c7bc-4fe4-9c24-ca0070a0775c\") " pod="openshift-machine-config-operator/machine-config-server-z7mtd" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729877 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgrvf\" (UniqueName: \"kubernetes.io/projected/de7a08c9-b456-475e-a688-56edc28603ac-kube-api-access-hgrvf\") pod \"package-server-manager-789f6589d5-hqnns\" (UID: \"de7a08c9-b456-475e-a688-56edc28603ac\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqnns" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729901 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/32f920f7-5adb-4cf3-bc1f-3a461b79bb10-images\") pod \"machine-config-operator-74547568cd-vrvxq\" (UID: \"32f920f7-5adb-4cf3-bc1f-3a461b79bb10\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729927 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3e2ce5bb-1abf-47c6-b6d2-59389118cff0-profile-collector-cert\") pod \"catalog-operator-68c6474976-btgsr\" (UID: \"3e2ce5bb-1abf-47c6-b6d2-59389118cff0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.729965 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzppq\" (UniqueName: \"kubernetes.io/projected/751d9838-4c34-4251-b3a6-1d808f98a034-kube-api-access-xzppq\") pod \"collect-profiles-29396445-v6dn4\" (UID: \"751d9838-4c34-4251-b3a6-1d808f98a034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.730176 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a33da900-6b7c-4157-8a6e-fde1cddf51ca-mountpoint-dir\") pod \"csi-hostpathplugin-8zp8h\" (UID: \"a33da900-6b7c-4157-8a6e-fde1cddf51ca\") " pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.730339 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a33da900-6b7c-4157-8a6e-fde1cddf51ca-registration-dir\") pod \"csi-hostpathplugin-8zp8h\" (UID: \"a33da900-6b7c-4157-8a6e-fde1cddf51ca\") " pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.730415 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a33da900-6b7c-4157-8a6e-fde1cddf51ca-plugins-dir\") pod \"csi-hostpathplugin-8zp8h\" (UID: \"a33da900-6b7c-4157-8a6e-fde1cddf51ca\") " pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.730481 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a33da900-6b7c-4157-8a6e-fde1cddf51ca-csi-data-dir\") pod \"csi-hostpathplugin-8zp8h\" (UID: \"a33da900-6b7c-4157-8a6e-fde1cddf51ca\") " pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.730882 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/215b4f1f-75ab-4c5c-aca8-26ea0edf5978-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rr7gt\" (UID: \"215b4f1f-75ab-4c5c-aca8-26ea0edf5978\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rr7gt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.745172 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.785245 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.804899 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.825007 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.845428 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.855057 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/215b4f1f-75ab-4c5c-aca8-26ea0edf5978-serving-cert\") pod \"openshift-config-operator-7777fb866f-rr7gt\" (UID: \"215b4f1f-75ab-4c5c-aca8-26ea0edf5978\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rr7gt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.864774 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.885336 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.898606 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2a1ecdb-4740-43c4-ba73-19d1f15465e1-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rsc75\" (UID: \"f2a1ecdb-4740-43c4-ba73-19d1f15465e1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rsc75" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.905693 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.926475 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.945145 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.953938 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2a1ecdb-4740-43c4-ba73-19d1f15465e1-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rsc75\" (UID: \"f2a1ecdb-4740-43c4-ba73-19d1f15465e1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rsc75" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.965218 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.977972 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/40771f13-f771-4e07-8ad9-668a3e48fb3b-etcd-client\") pod \"etcd-operator-b45778765-dhhh4\" (UID: \"40771f13-f771-4e07-8ad9-668a3e48fb3b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" Nov 22 04:49:54 crc kubenswrapper[4818]: I1122 04:49:54.985333 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.005165 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.025037 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.044782 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.053842 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40771f13-f771-4e07-8ad9-668a3e48fb3b-config\") pod \"etcd-operator-b45778765-dhhh4\" (UID: \"40771f13-f771-4e07-8ad9-668a3e48fb3b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.065622 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.075241 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40771f13-f771-4e07-8ad9-668a3e48fb3b-serving-cert\") pod \"etcd-operator-b45778765-dhhh4\" (UID: \"40771f13-f771-4e07-8ad9-668a3e48fb3b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.086378 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.095482 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/40771f13-f771-4e07-8ad9-668a3e48fb3b-etcd-ca\") pod \"etcd-operator-b45778765-dhhh4\" (UID: \"40771f13-f771-4e07-8ad9-668a3e48fb3b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.106848 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.117999 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/40771f13-f771-4e07-8ad9-668a3e48fb3b-etcd-service-ca\") pod \"etcd-operator-b45778765-dhhh4\" (UID: \"40771f13-f771-4e07-8ad9-668a3e48fb3b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.125095 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.145548 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.165133 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.172292 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea-proxy-tls\") pod \"machine-config-controller-84d6567774-4pxtx\" (UID: \"54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4pxtx" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.184906 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.205816 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.216553 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4aa421fd-173a-4c0d-8c66-ddcb7b173808-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jtg8k\" (UID: \"4aa421fd-173a-4c0d-8c66-ddcb7b173808\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jtg8k" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.224744 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.231726 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4aa421fd-173a-4c0d-8c66-ddcb7b173808-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jtg8k\" (UID: \"4aa421fd-173a-4c0d-8c66-ddcb7b173808\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jtg8k" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.245725 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.264544 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.285660 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.304556 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.324745 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.345849 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.350868 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/32f920f7-5adb-4cf3-bc1f-3a461b79bb10-proxy-tls\") pod \"machine-config-operator-74547568cd-vrvxq\" (UID: \"32f920f7-5adb-4cf3-bc1f-3a461b79bb10\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.365958 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.372628 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/32f920f7-5adb-4cf3-bc1f-3a461b79bb10-images\") pod \"machine-config-operator-74547568cd-vrvxq\" (UID: \"32f920f7-5adb-4cf3-bc1f-3a461b79bb10\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.384828 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.405028 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.411213 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1d4ef859-1409-451a-a2e1-4c1c8c6d473b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nz99w\" (UID: \"1d4ef859-1409-451a-a2e1-4c1c8c6d473b\") " pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.435773 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.442498 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1d4ef859-1409-451a-a2e1-4c1c8c6d473b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nz99w\" (UID: \"1d4ef859-1409-451a-a2e1-4c1c8c6d473b\") " pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.445633 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.465720 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.482729 4818 request.go:700] Waited for 1.015197216s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/secrets?fieldSelector=metadata.name%3Drouter-stats-default&limit=500&resourceVersion=0 Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.484975 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.498842 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/eb0cffc3-bf07-411c-822e-aa45c58387cb-stats-auth\") pod \"router-default-5444994796-5gfmz\" (UID: \"eb0cffc3-bf07-411c-822e-aa45c58387cb\") " pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.504839 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.525455 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.544971 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.555776 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb0cffc3-bf07-411c-822e-aa45c58387cb-metrics-certs\") pod \"router-default-5444994796-5gfmz\" (UID: \"eb0cffc3-bf07-411c-822e-aa45c58387cb\") " pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.565114 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.576589 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/eb0cffc3-bf07-411c-822e-aa45c58387cb-default-certificate\") pod \"router-default-5444994796-5gfmz\" (UID: \"eb0cffc3-bf07-411c-822e-aa45c58387cb\") " pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.590579 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.604707 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.606063 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eb0cffc3-bf07-411c-822e-aa45c58387cb-service-ca-bundle\") pod \"router-default-5444994796-5gfmz\" (UID: \"eb0cffc3-bf07-411c-822e-aa45c58387cb\") " pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.622642 4818 secret.go:188] Couldn't get secret openshift-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.622721 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5484670-b8e2-40b9-a36b-3fb47bb89421-serving-cert podName:d5484670-b8e2-40b9-a36b-3fb47bb89421 nodeName:}" failed. No retries permitted until 2025-11-22 04:49:56.122696293 +0000 UTC m=+148.697112850 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/d5484670-b8e2-40b9-a36b-3fb47bb89421-serving-cert") pod "controller-manager-879f6c89f-tjh65" (UID: "d5484670-b8e2-40b9-a36b-3fb47bb89421") : failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.623585 4818 configmap.go:193] Couldn't get configMap openshift-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.623700 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-config podName:d5484670-b8e2-40b9-a36b-3fb47bb89421 nodeName:}" failed. No retries permitted until 2025-11-22 04:49:56.12367005 +0000 UTC m=+148.698086677 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-config") pod "controller-manager-879f6c89f-tjh65" (UID: "d5484670-b8e2-40b9-a36b-3fb47bb89421") : failed to sync configmap cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.625646 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.625976 4818 configmap.go:193] Couldn't get configMap openshift-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.626058 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-client-ca podName:d5484670-b8e2-40b9-a36b-3fb47bb89421 nodeName:}" failed. No retries permitted until 2025-11-22 04:49:56.126036614 +0000 UTC m=+148.700453171 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-client-ca") pod "controller-manager-879f6c89f-tjh65" (UID: "d5484670-b8e2-40b9-a36b-3fb47bb89421") : failed to sync configmap cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.629995 4818 configmap.go:193] Couldn't get configMap openshift-machine-api/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.630117 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8d3b850f-8917-46af-b713-891a7c979d2e-config podName:8d3b850f-8917-46af-b713-891a7c979d2e nodeName:}" failed. No retries permitted until 2025-11-22 04:49:56.130089315 +0000 UTC m=+148.704505882 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/8d3b850f-8917-46af-b713-891a7c979d2e-config") pod "machine-api-operator-5694c8668f-8kzxl" (UID: "8d3b850f-8917-46af-b713-891a7c979d2e") : failed to sync configmap cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.645187 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.665059 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.672203 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/eb6893d2-46c1-46fa-8c40-5bbd900d9fc1-signing-cabundle\") pod \"service-ca-9c57cc56f-jp5fs\" (UID: \"eb6893d2-46c1-46fa-8c40-5bbd900d9fc1\") " pod="openshift-service-ca/service-ca-9c57cc56f-jp5fs" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.685908 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.704967 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.720714 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/eb6893d2-46c1-46fa-8c40-5bbd900d9fc1-signing-key\") pod \"service-ca-9c57cc56f-jp5fs\" (UID: \"eb6893d2-46c1-46fa-8c40-5bbd900d9fc1\") " pod="openshift-service-ca/service-ca-9c57cc56f-jp5fs" Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.723080 4818 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.723152 4818 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.723192 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a-config podName:b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a nodeName:}" failed. No retries permitted until 2025-11-22 04:49:56.223136873 +0000 UTC m=+148.797553440 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a-config") pod "service-ca-operator-777779d784-fxwd8" (UID: "b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a") : failed to sync configmap cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.723289 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/751d9838-4c34-4251-b3a6-1d808f98a034-secret-volume podName:751d9838-4c34-4251-b3a6-1d808f98a034 nodeName:}" failed. No retries permitted until 2025-11-22 04:49:56.223222465 +0000 UTC m=+148.797639032 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-volume" (UniqueName: "kubernetes.io/secret/751d9838-4c34-4251-b3a6-1d808f98a034-secret-volume") pod "collect-profiles-29396445-v6dn4" (UID: "751d9838-4c34-4251-b3a6-1d808f98a034") : failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.724896 4818 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.724988 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de7a08c9-b456-475e-a688-56edc28603ac-package-server-manager-serving-cert podName:de7a08c9-b456-475e-a688-56edc28603ac nodeName:}" failed. No retries permitted until 2025-11-22 04:49:56.224969042 +0000 UTC m=+148.799385609 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/de7a08c9-b456-475e-a688-56edc28603ac-package-server-manager-serving-cert") pod "package-server-manager-789f6589d5-hqnns" (UID: "de7a08c9-b456-475e-a688-56edc28603ac") : failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.725194 4818 secret.go:188] Couldn't get secret openshift-machine-api/control-plane-machine-set-operator-tls: failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.725409 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.725443 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7c429c72-1ae2-4b7d-9786-57ed9341bd48-control-plane-machine-set-operator-tls podName:7c429c72-1ae2-4b7d-9786-57ed9341bd48 nodeName:}" failed. No retries permitted until 2025-11-22 04:49:56.225376604 +0000 UTC m=+148.799793161 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" (UniqueName: "kubernetes.io/secret/7c429c72-1ae2-4b7d-9786-57ed9341bd48-control-plane-machine-set-operator-tls") pod "control-plane-machine-set-operator-78cbb6b69f-zftfz" (UID: "7c429c72-1ae2-4b7d-9786-57ed9341bd48") : failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.725680 4818 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.725887 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/751d9838-4c34-4251-b3a6-1d808f98a034-config-volume podName:751d9838-4c34-4251-b3a6-1d808f98a034 nodeName:}" failed. No retries permitted until 2025-11-22 04:49:56.225836616 +0000 UTC m=+148.800253183 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/751d9838-4c34-4251-b3a6-1d808f98a034-config-volume") pod "collect-profiles-29396445-v6dn4" (UID: "751d9838-4c34-4251-b3a6-1d808f98a034") : failed to sync configmap cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.726775 4818 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.726849 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe-profile-collector-cert podName:5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe nodeName:}" failed. No retries permitted until 2025-11-22 04:49:56.226827163 +0000 UTC m=+148.801243720 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe-profile-collector-cert") pod "olm-operator-6b444d44fb-jlvsb" (UID: "5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe") : failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.730466 4818 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.730563 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9472804-f8b1-440a-b12e-d351630ed795-apiservice-cert podName:c9472804-f8b1-440a-b12e-d351630ed795 nodeName:}" failed. No retries permitted until 2025-11-22 04:49:56.230538463 +0000 UTC m=+148.804955030 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/c9472804-f8b1-440a-b12e-d351630ed795-apiservice-cert") pod "packageserver-d55dfcdfc-k27q8" (UID: "c9472804-f8b1-440a-b12e-d351630ed795") : failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.730610 4818 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.730653 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f239350-c7bc-4fe4-9c24-ca0070a0775c-node-bootstrap-token podName:6f239350-c7bc-4fe4-9c24-ca0070a0775c nodeName:}" failed. No retries permitted until 2025-11-22 04:49:56.230640176 +0000 UTC m=+148.805056743 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/6f239350-c7bc-4fe4-9c24-ca0070a0775c-node-bootstrap-token") pod "machine-config-server-z7mtd" (UID: "6f239350-c7bc-4fe4-9c24-ca0070a0775c") : failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.730685 4818 secret.go:188] Couldn't get secret openshift-multus/multus-admission-controller-secret: failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.730741 4818 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.730702 4818 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.730834 4818 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.730808 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3e2ce5bb-1abf-47c6-b6d2-59389118cff0-profile-collector-cert podName:3e2ce5bb-1abf-47c6-b6d2-59389118cff0 nodeName:}" failed. No retries permitted until 2025-11-22 04:49:56.230790441 +0000 UTC m=+148.805206998 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/3e2ce5bb-1abf-47c6-b6d2-59389118cff0-profile-collector-cert") pod "catalog-operator-68c6474976-btgsr" (UID: "3e2ce5bb-1abf-47c6-b6d2-59389118cff0") : failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.730877 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7ab520a0-54e7-4923-867c-215fffacf270-webhook-certs podName:7ab520a0-54e7-4923-867c-215fffacf270 nodeName:}" failed. No retries permitted until 2025-11-22 04:49:56.230858893 +0000 UTC m=+148.805275470 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/7ab520a0-54e7-4923-867c-215fffacf270-webhook-certs") pod "multus-admission-controller-857f4d67dd-7bmhw" (UID: "7ab520a0-54e7-4923-867c-215fffacf270") : failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.730879 4818 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.730900 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f239350-c7bc-4fe4-9c24-ca0070a0775c-certs podName:6f239350-c7bc-4fe4-9c24-ca0070a0775c nodeName:}" failed. No retries permitted until 2025-11-22 04:49:56.230888794 +0000 UTC m=+148.805305361 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/6f239350-c7bc-4fe4-9c24-ca0070a0775c-certs") pod "machine-config-server-z7mtd" (UID: "6f239350-c7bc-4fe4-9c24-ca0070a0775c") : failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.730924 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3e2ce5bb-1abf-47c6-b6d2-59389118cff0-srv-cert podName:3e2ce5bb-1abf-47c6-b6d2-59389118cff0 nodeName:}" failed. No retries permitted until 2025-11-22 04:49:56.230912594 +0000 UTC m=+148.805329161 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/3e2ce5bb-1abf-47c6-b6d2-59389118cff0-srv-cert") pod "catalog-operator-68c6474976-btgsr" (UID: "3e2ce5bb-1abf-47c6-b6d2-59389118cff0") : failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.730947 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2fde1815-9a0b-4f85-820c-b482ccdeab3c-cert podName:2fde1815-9a0b-4f85-820c-b482ccdeab3c nodeName:}" failed. No retries permitted until 2025-11-22 04:49:56.230936425 +0000 UTC m=+148.805352992 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2fde1815-9a0b-4f85-820c-b482ccdeab3c-cert") pod "ingress-canary-ccd5x" (UID: "2fde1815-9a0b-4f85-820c-b482ccdeab3c") : failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.730969 4818 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.730980 4818 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.730963 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a-serving-cert\") pod \"service-ca-operator-777779d784-fxwd8\" (UID: \"b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fxwd8" Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.731025 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9472804-f8b1-440a-b12e-d351630ed795-webhook-cert podName:c9472804-f8b1-440a-b12e-d351630ed795 nodeName:}" failed. No retries permitted until 2025-11-22 04:49:56.231014107 +0000 UTC m=+148.805430674 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/c9472804-f8b1-440a-b12e-d351630ed795-webhook-cert") pod "packageserver-d55dfcdfc-k27q8" (UID: "c9472804-f8b1-440a-b12e-d351630ed795") : failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: E1122 04:49:55.731054 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe-srv-cert podName:5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe nodeName:}" failed. No retries permitted until 2025-11-22 04:49:56.231040228 +0000 UTC m=+148.805456795 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe-srv-cert") pod "olm-operator-6b444d44fb-jlvsb" (UID: "5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe") : failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.745679 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.765330 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.785386 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.805093 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.825571 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.845306 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.864724 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.885924 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.905509 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.924894 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.945467 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.964877 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 22 04:49:55 crc kubenswrapper[4818]: I1122 04:49:55.985030 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.005662 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.026066 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.045581 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.065054 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.085320 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.106123 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.125001 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.145071 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.159822 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-client-ca\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.159973 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5484670-b8e2-40b9-a36b-3fb47bb89421-serving-cert\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.160316 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d3b850f-8917-46af-b713-891a7c979d2e-config\") pod \"machine-api-operator-5694c8668f-8kzxl\" (UID: \"8d3b850f-8917-46af-b713-891a7c979d2e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8kzxl" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.160424 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-config\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.165655 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.185174 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.210191 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.226445 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.246580 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.261467 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe-profile-collector-cert\") pod \"olm-operator-6b444d44fb-jlvsb\" (UID: \"5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.261585 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c9472804-f8b1-440a-b12e-d351630ed795-apiservice-cert\") pod \"packageserver-d55dfcdfc-k27q8\" (UID: \"c9472804-f8b1-440a-b12e-d351630ed795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.261627 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6f239350-c7bc-4fe4-9c24-ca0070a0775c-certs\") pod \"machine-config-server-z7mtd\" (UID: \"6f239350-c7bc-4fe4-9c24-ca0070a0775c\") " pod="openshift-machine-config-operator/machine-config-server-z7mtd" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.261711 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3e2ce5bb-1abf-47c6-b6d2-59389118cff0-srv-cert\") pod \"catalog-operator-68c6474976-btgsr\" (UID: \"3e2ce5bb-1abf-47c6-b6d2-59389118cff0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.261769 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2fde1815-9a0b-4f85-820c-b482ccdeab3c-cert\") pod \"ingress-canary-ccd5x\" (UID: \"2fde1815-9a0b-4f85-820c-b482ccdeab3c\") " pod="openshift-ingress-canary/ingress-canary-ccd5x" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.262064 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe-srv-cert\") pod \"olm-operator-6b444d44fb-jlvsb\" (UID: \"5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.262143 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7ab520a0-54e7-4923-867c-215fffacf270-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7bmhw\" (UID: \"7ab520a0-54e7-4923-867c-215fffacf270\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7bmhw" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.262278 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c9472804-f8b1-440a-b12e-d351630ed795-webhook-cert\") pod \"packageserver-d55dfcdfc-k27q8\" (UID: \"c9472804-f8b1-440a-b12e-d351630ed795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.262340 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6f239350-c7bc-4fe4-9c24-ca0070a0775c-node-bootstrap-token\") pod \"machine-config-server-z7mtd\" (UID: \"6f239350-c7bc-4fe4-9c24-ca0070a0775c\") " pod="openshift-machine-config-operator/machine-config-server-z7mtd" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.262400 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3e2ce5bb-1abf-47c6-b6d2-59389118cff0-profile-collector-cert\") pod \"catalog-operator-68c6474976-btgsr\" (UID: \"3e2ce5bb-1abf-47c6-b6d2-59389118cff0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.262566 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/751d9838-4c34-4251-b3a6-1d808f98a034-secret-volume\") pod \"collect-profiles-29396445-v6dn4\" (UID: \"751d9838-4c34-4251-b3a6-1d808f98a034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.262612 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a-config\") pod \"service-ca-operator-777779d784-fxwd8\" (UID: \"b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fxwd8" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.262685 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7c429c72-1ae2-4b7d-9786-57ed9341bd48-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zftfz\" (UID: \"7c429c72-1ae2-4b7d-9786-57ed9341bd48\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zftfz" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.262735 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/de7a08c9-b456-475e-a688-56edc28603ac-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hqnns\" (UID: \"de7a08c9-b456-475e-a688-56edc28603ac\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqnns" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.262875 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/751d9838-4c34-4251-b3a6-1d808f98a034-config-volume\") pod \"collect-profiles-29396445-v6dn4\" (UID: \"751d9838-4c34-4251-b3a6-1d808f98a034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.265017 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/751d9838-4c34-4251-b3a6-1d808f98a034-config-volume\") pod \"collect-profiles-29396445-v6dn4\" (UID: \"751d9838-4c34-4251-b3a6-1d808f98a034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.266941 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3e2ce5bb-1abf-47c6-b6d2-59389118cff0-srv-cert\") pod \"catalog-operator-68c6474976-btgsr\" (UID: \"3e2ce5bb-1abf-47c6-b6d2-59389118cff0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.268450 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c9472804-f8b1-440a-b12e-d351630ed795-apiservice-cert\") pod \"packageserver-d55dfcdfc-k27q8\" (UID: \"c9472804-f8b1-440a-b12e-d351630ed795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.268930 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.270445 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c9472804-f8b1-440a-b12e-d351630ed795-webhook-cert\") pod \"packageserver-d55dfcdfc-k27q8\" (UID: \"c9472804-f8b1-440a-b12e-d351630ed795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.270458 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a-config\") pod \"service-ca-operator-777779d784-fxwd8\" (UID: \"b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fxwd8" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.271825 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7c429c72-1ae2-4b7d-9786-57ed9341bd48-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zftfz\" (UID: \"7c429c72-1ae2-4b7d-9786-57ed9341bd48\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zftfz" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.273883 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/de7a08c9-b456-475e-a688-56edc28603ac-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hqnns\" (UID: \"de7a08c9-b456-475e-a688-56edc28603ac\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqnns" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.274630 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe-profile-collector-cert\") pod \"olm-operator-6b444d44fb-jlvsb\" (UID: \"5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.274828 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/751d9838-4c34-4251-b3a6-1d808f98a034-secret-volume\") pod \"collect-profiles-29396445-v6dn4\" (UID: \"751d9838-4c34-4251-b3a6-1d808f98a034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.275339 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2fde1815-9a0b-4f85-820c-b482ccdeab3c-cert\") pod \"ingress-canary-ccd5x\" (UID: \"2fde1815-9a0b-4f85-820c-b482ccdeab3c\") " pod="openshift-ingress-canary/ingress-canary-ccd5x" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.276197 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe-srv-cert\") pod \"olm-operator-6b444d44fb-jlvsb\" (UID: \"5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.276995 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3e2ce5bb-1abf-47c6-b6d2-59389118cff0-profile-collector-cert\") pod \"catalog-operator-68c6474976-btgsr\" (UID: \"3e2ce5bb-1abf-47c6-b6d2-59389118cff0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.278845 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7ab520a0-54e7-4923-867c-215fffacf270-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7bmhw\" (UID: \"7ab520a0-54e7-4923-867c-215fffacf270\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7bmhw" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.305218 4818 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.325120 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.344836 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.365510 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.414093 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9zfh\" (UniqueName: \"kubernetes.io/projected/d7c47e41-70c7-4784-aa37-c609650e3e1b-kube-api-access-h9zfh\") pod \"migrator-59844c95c7-q6xbv\" (UID: \"d7c47e41-70c7-4784-aa37-c609650e3e1b\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q6xbv" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.433011 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6f239350-c7bc-4fe4-9c24-ca0070a0775c-node-bootstrap-token\") pod \"machine-config-server-z7mtd\" (UID: \"6f239350-c7bc-4fe4-9c24-ca0070a0775c\") " pod="openshift-machine-config-operator/machine-config-server-z7mtd" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.434632 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6f239350-c7bc-4fe4-9c24-ca0070a0775c-certs\") pod \"machine-config-server-z7mtd\" (UID: \"6f239350-c7bc-4fe4-9c24-ca0070a0775c\") " pod="openshift-machine-config-operator/machine-config-server-z7mtd" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.449968 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87hd4\" (UniqueName: \"kubernetes.io/projected/e48b2f0e-a02f-48c6-b513-33537b0108b6-kube-api-access-87hd4\") pod \"openshift-apiserver-operator-796bbdcf4f-xvd5r\" (UID: \"e48b2f0e-a02f-48c6-b513-33537b0108b6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xvd5r" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.471491 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjkhp\" (UniqueName: \"kubernetes.io/projected/0dcaa80a-8993-4452-9acb-b8010eade7c6-kube-api-access-sjkhp\") pod \"oauth-openshift-558db77b4-mkgj4\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.483106 4818 request.go:700] Waited for 1.862973114s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-oauth-apiserver/serviceaccounts/oauth-apiserver-sa/token Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.491968 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgpx6\" (UniqueName: \"kubernetes.io/projected/5a66393b-1460-4ef7-9c8a-2001a0d187fb-kube-api-access-jgpx6\") pod \"authentication-operator-69f744f599-s8xvt\" (UID: \"5a66393b-1460-4ef7-9c8a-2001a0d187fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8xvt" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.510968 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md47q\" (UniqueName: \"kubernetes.io/projected/10918dd9-6412-4172-a0f3-eac3fbd10da1-kube-api-access-md47q\") pod \"apiserver-7bbb656c7d-4lh8z\" (UID: \"10918dd9-6412-4172-a0f3-eac3fbd10da1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.529786 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1c05f6d6-e233-46ca-91df-eaba2614018f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-92bz6\" (UID: \"1c05f6d6-e233-46ca-91df-eaba2614018f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.533946 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.551908 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvx6l\" (UniqueName: \"kubernetes.io/projected/85928d04-1176-4b1d-8bb7-d55376576225-kube-api-access-mvx6l\") pod \"cluster-image-registry-operator-dc59b4c8b-2ps29\" (UID: \"85928d04-1176-4b1d-8bb7-d55376576225\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ps29" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.574426 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwpwz\" (UniqueName: \"kubernetes.io/projected/7337e50a-daf8-4f50-9060-ec2a14752444-kube-api-access-kwpwz\") pod \"console-operator-58897d9998-44grv\" (UID: \"7337e50a-daf8-4f50-9060-ec2a14752444\") " pod="openshift-console-operator/console-operator-58897d9998-44grv" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.584782 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-s8xvt" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.610999 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95th8\" (UniqueName: \"kubernetes.io/projected/61feab0c-5f50-49fe-bf35-de2a5f66bf32-kube-api-access-95th8\") pod \"downloads-7954f5f757-rcvlq\" (UID: \"61feab0c-5f50-49fe-bf35-de2a5f66bf32\") " pod="openshift-console/downloads-7954f5f757-rcvlq" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.614961 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/45ba8e80-12bf-472e-a0b7-b25601a9074e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-887w5\" (UID: \"45ba8e80-12bf-472e-a0b7-b25601a9074e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-887w5" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.623127 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.639353 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/85928d04-1176-4b1d-8bb7-d55376576225-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2ps29\" (UID: \"85928d04-1176-4b1d-8bb7-d55376576225\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ps29" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.639626 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xvd5r" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.640162 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-rcvlq" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.649397 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j7fh\" (UniqueName: \"kubernetes.io/projected/670a3117-320e-477c-8c4b-a54fe5a4ae9c-kube-api-access-8j7fh\") pod \"console-f9d7485db-h56w2\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.651459 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-44grv" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.664359 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-887w5" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.671101 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/312c389c-57c0-4709-9320-777e4659d499-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-28f5t\" (UID: \"312c389c-57c0-4709-9320-777e4659d499\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-28f5t" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.685367 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-28f5t" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.697321 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ps29" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.697514 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gsdk\" (UniqueName: \"kubernetes.io/projected/1c05f6d6-e233-46ca-91df-eaba2614018f-kube-api-access-8gsdk\") pod \"ingress-operator-5b745b69d9-92bz6\" (UID: \"1c05f6d6-e233-46ca-91df-eaba2614018f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.705392 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.714705 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q6xbv" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.723050 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdkln\" (UniqueName: \"kubernetes.io/projected/968b0694-9a30-49ca-bee7-d1eadbb31505-kube-api-access-jdkln\") pod \"machine-approver-56656f9798-h7srs\" (UID: \"968b0694-9a30-49ca-bee7-d1eadbb31505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h7srs" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.725741 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs2s8\" (UniqueName: \"kubernetes.io/projected/62bdffb2-a1c7-4960-81aa-b570834f78cd-kube-api-access-bs2s8\") pod \"cluster-samples-operator-665b6dd947-j7ttg\" (UID: \"62bdffb2-a1c7-4960-81aa-b570834f78cd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j7ttg" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.745321 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtpcw\" (UniqueName: \"kubernetes.io/projected/4bfeb5de-dcfb-4f0b-b118-112ee3262c18-kube-api-access-qtpcw\") pod \"dns-operator-744455d44c-dvt5z\" (UID: \"4bfeb5de-dcfb-4f0b-b118-112ee3262c18\") " pod="openshift-dns-operator/dns-operator-744455d44c-dvt5z" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.762632 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcll2\" (UniqueName: \"kubernetes.io/projected/8d3b850f-8917-46af-b713-891a7c979d2e-kube-api-access-tcll2\") pod \"machine-api-operator-5694c8668f-8kzxl\" (UID: \"8d3b850f-8917-46af-b713-891a7c979d2e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8kzxl" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.782039 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23111038-ec66-4295-a6eb-b847c52a0c81-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7f58c\" (UID: \"23111038-ec66-4295-a6eb-b847c52a0c81\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7f58c" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.802009 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq7dc\" (UniqueName: \"kubernetes.io/projected/091bb385-63c1-443f-9aec-bedd08514ee8-kube-api-access-jq7dc\") pod \"apiserver-76f77b778f-9nsph\" (UID: \"091bb385-63c1-443f-9aec-bedd08514ee8\") " pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.824588 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqrct\" (UniqueName: \"kubernetes.io/projected/353ed841-0d3b-46df-beec-1906f6d801fd-kube-api-access-fqrct\") pod \"route-controller-manager-6576b87f9c-4lz4x\" (UID: \"353ed841-0d3b-46df-beec-1906f6d801fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.847851 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk6gh\" (UniqueName: \"kubernetes.io/projected/a33da900-6b7c-4157-8a6e-fde1cddf51ca-kube-api-access-zk6gh\") pod \"csi-hostpathplugin-8zp8h\" (UID: \"a33da900-6b7c-4157-8a6e-fde1cddf51ca\") " pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.852083 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z"] Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.858696 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpq48\" (UniqueName: \"kubernetes.io/projected/1d4ef859-1409-451a-a2e1-4c1c8c6d473b-kube-api-access-kpq48\") pod \"marketplace-operator-79b997595-nz99w\" (UID: \"1d4ef859-1409-451a-a2e1-4c1c8c6d473b\") " pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.867061 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.872562 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.882876 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cccc9\" (UniqueName: \"kubernetes.io/projected/f2a1ecdb-4740-43c4-ba73-19d1f15465e1-kube-api-access-cccc9\") pod \"openshift-controller-manager-operator-756b6f6bc6-rsc75\" (UID: \"f2a1ecdb-4740-43c4-ba73-19d1f15465e1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rsc75" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.900441 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h7srs" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.909724 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77sqd\" (UniqueName: \"kubernetes.io/projected/b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a-kube-api-access-77sqd\") pod \"service-ca-operator-777779d784-fxwd8\" (UID: \"b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fxwd8" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.916550 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j7ttg" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.919839 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9qtj\" (UniqueName: \"kubernetes.io/projected/2fde1815-9a0b-4f85-820c-b482ccdeab3c-kube-api-access-g9qtj\") pod \"ingress-canary-ccd5x\" (UID: \"2fde1815-9a0b-4f85-820c-b482ccdeab3c\") " pod="openshift-ingress-canary/ingress-canary-ccd5x" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.920049 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-ccd5x" Nov 22 04:49:56 crc kubenswrapper[4818]: W1122 04:49:56.930661 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod968b0694_9a30_49ca_bee7_d1eadbb31505.slice/crio-86a81a0d2b727eb6cbd5a966f1ae492578ddb0855fec946c047354d849a9c27d WatchSource:0}: Error finding container 86a81a0d2b727eb6cbd5a966f1ae492578ddb0855fec946c047354d849a9c27d: Status 404 returned error can't find the container with id 86a81a0d2b727eb6cbd5a966f1ae492578ddb0855fec946c047354d849a9c27d Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.940870 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.944912 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66fgc\" (UniqueName: \"kubernetes.io/projected/4aa421fd-173a-4c0d-8c66-ddcb7b173808-kube-api-access-66fgc\") pod \"kube-storage-version-migrator-operator-b67b599dd-jtg8k\" (UID: \"4aa421fd-173a-4c0d-8c66-ddcb7b173808\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jtg8k" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.957701 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-dvt5z" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.959345 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br4mq\" (UniqueName: \"kubernetes.io/projected/eb6893d2-46c1-46fa-8c40-5bbd900d9fc1-kube-api-access-br4mq\") pod \"service-ca-9c57cc56f-jp5fs\" (UID: \"eb6893d2-46c1-46fa-8c40-5bbd900d9fc1\") " pod="openshift-service-ca/service-ca-9c57cc56f-jp5fs" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.972592 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.980416 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6s8p\" (UniqueName: \"kubernetes.io/projected/7c429c72-1ae2-4b7d-9786-57ed9341bd48-kube-api-access-n6s8p\") pod \"control-plane-machine-set-operator-78cbb6b69f-zftfz\" (UID: \"7c429c72-1ae2-4b7d-9786-57ed9341bd48\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zftfz" Nov 22 04:49:56 crc kubenswrapper[4818]: I1122 04:49:56.997658 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79g6t\" (UniqueName: \"kubernetes.io/projected/215b4f1f-75ab-4c5c-aca8-26ea0edf5978-kube-api-access-79g6t\") pod \"openshift-config-operator-7777fb866f-rr7gt\" (UID: \"215b4f1f-75ab-4c5c-aca8-26ea0edf5978\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rr7gt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.001467 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7f58c" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.021756 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rr7gt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.022974 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ns7n\" (UniqueName: \"kubernetes.io/projected/7ab520a0-54e7-4923-867c-215fffacf270-kube-api-access-2ns7n\") pod \"multus-admission-controller-857f4d67dd-7bmhw\" (UID: \"7ab520a0-54e7-4923-867c-215fffacf270\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7bmhw" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.049484 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thfxc\" (UniqueName: \"kubernetes.io/projected/3e2ce5bb-1abf-47c6-b6d2-59389118cff0-kube-api-access-thfxc\") pod \"catalog-operator-68c6474976-btgsr\" (UID: \"3e2ce5bb-1abf-47c6-b6d2-59389118cff0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.063374 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhqfr\" (UniqueName: \"kubernetes.io/projected/40771f13-f771-4e07-8ad9-668a3e48fb3b-kube-api-access-qhqfr\") pod \"etcd-operator-b45778765-dhhh4\" (UID: \"40771f13-f771-4e07-8ad9-668a3e48fb3b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.074793 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rsc75" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.077769 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.079161 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzppq\" (UniqueName: \"kubernetes.io/projected/751d9838-4c34-4251-b3a6-1d808f98a034-kube-api-access-xzppq\") pod \"collect-profiles-29396445-v6dn4\" (UID: \"751d9838-4c34-4251-b3a6-1d808f98a034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.100859 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jtg8k" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.106242 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v28zk\" (UniqueName: \"kubernetes.io/projected/c9472804-f8b1-440a-b12e-d351630ed795-kube-api-access-v28zk\") pod \"packageserver-d55dfcdfc-k27q8\" (UID: \"c9472804-f8b1-440a-b12e-d351630ed795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.121727 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8vfr\" (UniqueName: \"kubernetes.io/projected/5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe-kube-api-access-j8vfr\") pod \"olm-operator-6b444d44fb-jlvsb\" (UID: \"5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.123498 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.144339 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fpr4\" (UniqueName: \"kubernetes.io/projected/eb0cffc3-bf07-411c-822e-aa45c58387cb-kube-api-access-8fpr4\") pod \"router-default-5444994796-5gfmz\" (UID: \"eb0cffc3-bf07-411c-822e-aa45c58387cb\") " pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.147436 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-jp5fs" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.147767 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-rcvlq"] Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.151479 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.153451 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-s8xvt"] Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.158218 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fxwd8" Nov 22 04:49:57 crc kubenswrapper[4818]: E1122 04:49:57.160558 4818 secret.go:188] Couldn't get secret openshift-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:57 crc kubenswrapper[4818]: E1122 04:49:57.160622 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5484670-b8e2-40b9-a36b-3fb47bb89421-serving-cert podName:d5484670-b8e2-40b9-a36b-3fb47bb89421 nodeName:}" failed. No retries permitted until 2025-11-22 04:49:58.160604804 +0000 UTC m=+150.735021331 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/d5484670-b8e2-40b9-a36b-3fb47bb89421-serving-cert") pod "controller-manager-879f6c89f-tjh65" (UID: "d5484670-b8e2-40b9-a36b-3fb47bb89421") : failed to sync secret cache: timed out waiting for the condition Nov 22 04:49:57 crc kubenswrapper[4818]: E1122 04:49:57.160664 4818 configmap.go:193] Couldn't get configMap openshift-machine-api/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Nov 22 04:49:57 crc kubenswrapper[4818]: E1122 04:49:57.160727 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8d3b850f-8917-46af-b713-891a7c979d2e-config podName:8d3b850f-8917-46af-b713-891a7c979d2e nodeName:}" failed. No retries permitted until 2025-11-22 04:49:58.160706327 +0000 UTC m=+150.735122854 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/8d3b850f-8917-46af-b713-891a7c979d2e-config") pod "machine-api-operator-5694c8668f-8kzxl" (UID: "8d3b850f-8917-46af-b713-891a7c979d2e") : failed to sync configmap cache: timed out waiting for the condition Nov 22 04:49:57 crc kubenswrapper[4818]: E1122 04:49:57.160758 4818 configmap.go:193] Couldn't get configMap openshift-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Nov 22 04:49:57 crc kubenswrapper[4818]: E1122 04:49:57.160780 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-config podName:d5484670-b8e2-40b9-a36b-3fb47bb89421 nodeName:}" failed. No retries permitted until 2025-11-22 04:49:58.160774349 +0000 UTC m=+150.735190876 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-config") pod "controller-manager-879f6c89f-tjh65" (UID: "d5484670-b8e2-40b9-a36b-3fb47bb89421") : failed to sync configmap cache: timed out waiting for the condition Nov 22 04:49:57 crc kubenswrapper[4818]: E1122 04:49:57.160844 4818 configmap.go:193] Couldn't get configMap openshift-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Nov 22 04:49:57 crc kubenswrapper[4818]: E1122 04:49:57.161014 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-client-ca podName:d5484670-b8e2-40b9-a36b-3fb47bb89421 nodeName:}" failed. No retries permitted until 2025-11-22 04:49:58.160971534 +0000 UTC m=+150.735388071 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-client-ca") pod "controller-manager-879f6c89f-tjh65" (UID: "d5484670-b8e2-40b9-a36b-3fb47bb89421") : failed to sync configmap cache: timed out waiting for the condition Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.163755 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtddb\" (UniqueName: \"kubernetes.io/projected/54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea-kube-api-access-wtddb\") pod \"machine-config-controller-84d6567774-4pxtx\" (UID: \"54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4pxtx" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.166469 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.166883 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h7srs" event={"ID":"968b0694-9a30-49ca-bee7-d1eadbb31505","Type":"ContainerStarted","Data":"86a81a0d2b727eb6cbd5a966f1ae492578ddb0855fec946c047354d849a9c27d"} Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.172762 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xvd5r"] Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.173633 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7bmhw" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.177000 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9r5z\" (UniqueName: \"kubernetes.io/projected/6f239350-c7bc-4fe4-9c24-ca0070a0775c-kube-api-access-x9r5z\") pod \"machine-config-server-z7mtd\" (UID: \"6f239350-c7bc-4fe4-9c24-ca0070a0775c\") " pod="openshift-machine-config-operator/machine-config-server-z7mtd" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.181553 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-z7mtd" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.187798 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" event={"ID":"10918dd9-6412-4172-a0f3-eac3fbd10da1","Type":"ContainerStarted","Data":"44dcd528b772c13b12b162ebc2eb5ca11e4a415b340efadc3a632cc595693152"} Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.194520 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.198317 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zftfz" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.203130 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgrvf\" (UniqueName: \"kubernetes.io/projected/de7a08c9-b456-475e-a688-56edc28603ac-kube-api-access-hgrvf\") pod \"package-server-manager-789f6589d5-hqnns\" (UID: \"de7a08c9-b456-475e-a688-56edc28603ac\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqnns" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.211751 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.231523 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vlnj\" (UniqueName: \"kubernetes.io/projected/32f920f7-5adb-4cf3-bc1f-3a461b79bb10-kube-api-access-7vlnj\") pod \"machine-config-operator-74547568cd-vrvxq\" (UID: \"32f920f7-5adb-4cf3-bc1f-3a461b79bb10\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.244545 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.248832 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-44grv"] Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.256625 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-28f5t"] Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.258401 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-h56w2"] Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.264616 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.274172 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mkgj4"] Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.285155 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.307317 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.328412 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58r46\" (UniqueName: \"kubernetes.io/projected/d5484670-b8e2-40b9-a36b-3fb47bb89421-kube-api-access-58r46\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.329498 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.346616 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.367551 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.369135 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ps29"] Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.370401 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-q6xbv"] Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.374017 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-887w5"] Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.385135 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.385222 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4pxtx" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.416424 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.433139 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.480305 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5fdv\" (UniqueName: \"kubernetes.io/projected/677aa266-f94d-498a-8034-1205889c4958-kube-api-access-n5fdv\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.480360 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.480384 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/677aa266-f94d-498a-8034-1205889c4958-bound-sa-token\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.480521 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/677aa266-f94d-498a-8034-1205889c4958-registry-certificates\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.480610 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/677aa266-f94d-498a-8034-1205889c4958-trusted-ca\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.480789 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/677aa266-f94d-498a-8034-1205889c4958-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.480834 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/677aa266-f94d-498a-8034-1205889c4958-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.480891 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/677aa266-f94d-498a-8034-1205889c4958-registry-tls\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: E1122 04:49:57.487408 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:49:57.987390944 +0000 UTC m=+150.561807471 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.501775 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqnns" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.579591 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-ccd5x"] Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.581696 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.581938 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ffee77a3-faea-4547-989b-02aab97c5471-metrics-tls\") pod \"dns-default-b86wj\" (UID: \"ffee77a3-faea-4547-989b-02aab97c5471\") " pod="openshift-dns/dns-default-b86wj" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.582116 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffee77a3-faea-4547-989b-02aab97c5471-config-volume\") pod \"dns-default-b86wj\" (UID: \"ffee77a3-faea-4547-989b-02aab97c5471\") " pod="openshift-dns/dns-default-b86wj" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.582147 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/677aa266-f94d-498a-8034-1205889c4958-registry-certificates\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.582341 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/677aa266-f94d-498a-8034-1205889c4958-trusted-ca\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.582548 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n87g4\" (UniqueName: \"kubernetes.io/projected/ffee77a3-faea-4547-989b-02aab97c5471-kube-api-access-n87g4\") pod \"dns-default-b86wj\" (UID: \"ffee77a3-faea-4547-989b-02aab97c5471\") " pod="openshift-dns/dns-default-b86wj" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.582608 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/677aa266-f94d-498a-8034-1205889c4958-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.582674 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/677aa266-f94d-498a-8034-1205889c4958-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.582737 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/677aa266-f94d-498a-8034-1205889c4958-registry-tls\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.582839 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5fdv\" (UniqueName: \"kubernetes.io/projected/677aa266-f94d-498a-8034-1205889c4958-kube-api-access-n5fdv\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.582967 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/677aa266-f94d-498a-8034-1205889c4958-bound-sa-token\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: E1122 04:49:57.583298 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:49:58.0832831 +0000 UTC m=+150.657699627 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.589116 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/677aa266-f94d-498a-8034-1205889c4958-registry-certificates\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.594868 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/677aa266-f94d-498a-8034-1205889c4958-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.597541 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-8zp8h"] Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.600386 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/677aa266-f94d-498a-8034-1205889c4958-trusted-ca\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.600782 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/677aa266-f94d-498a-8034-1205889c4958-registry-tls\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.600986 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-9nsph"] Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.602118 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6"] Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.607931 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/677aa266-f94d-498a-8034-1205889c4958-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.610397 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x"] Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.611833 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-dvt5z"] Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.679099 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5fdv\" (UniqueName: \"kubernetes.io/projected/677aa266-f94d-498a-8034-1205889c4958-kube-api-access-n5fdv\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.681908 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/677aa266-f94d-498a-8034-1205889c4958-bound-sa-token\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.686853 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.687064 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ffee77a3-faea-4547-989b-02aab97c5471-metrics-tls\") pod \"dns-default-b86wj\" (UID: \"ffee77a3-faea-4547-989b-02aab97c5471\") " pod="openshift-dns/dns-default-b86wj" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.687101 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffee77a3-faea-4547-989b-02aab97c5471-config-volume\") pod \"dns-default-b86wj\" (UID: \"ffee77a3-faea-4547-989b-02aab97c5471\") " pod="openshift-dns/dns-default-b86wj" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.687152 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n87g4\" (UniqueName: \"kubernetes.io/projected/ffee77a3-faea-4547-989b-02aab97c5471-kube-api-access-n87g4\") pod \"dns-default-b86wj\" (UID: \"ffee77a3-faea-4547-989b-02aab97c5471\") " pod="openshift-dns/dns-default-b86wj" Nov 22 04:49:57 crc kubenswrapper[4818]: E1122 04:49:57.687553 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:49:58.187542393 +0000 UTC m=+150.761958920 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.688353 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffee77a3-faea-4547-989b-02aab97c5471-config-volume\") pod \"dns-default-b86wj\" (UID: \"ffee77a3-faea-4547-989b-02aab97c5471\") " pod="openshift-dns/dns-default-b86wj" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.714482 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ffee77a3-faea-4547-989b-02aab97c5471-metrics-tls\") pod \"dns-default-b86wj\" (UID: \"ffee77a3-faea-4547-989b-02aab97c5471\") " pod="openshift-dns/dns-default-b86wj" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.725670 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n87g4\" (UniqueName: \"kubernetes.io/projected/ffee77a3-faea-4547-989b-02aab97c5471-kube-api-access-n87g4\") pod \"dns-default-b86wj\" (UID: \"ffee77a3-faea-4547-989b-02aab97c5471\") " pod="openshift-dns/dns-default-b86wj" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.761380 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j7ttg"] Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.761543 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rr7gt"] Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.780225 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7f58c"] Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.802617 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:49:57 crc kubenswrapper[4818]: E1122 04:49:57.802765 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:49:58.302738983 +0000 UTC m=+150.877155510 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.802852 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:57 crc kubenswrapper[4818]: E1122 04:49:57.803162 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:49:58.303151924 +0000 UTC m=+150.877568451 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:57 crc kubenswrapper[4818]: W1122 04:49:57.819451 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod091bb385_63c1_443f_9aec_bedd08514ee8.slice/crio-30c7015f702e43d2eec6457926f29905b242def9b2f9e423340ce74bd3ecf8a1 WatchSource:0}: Error finding container 30c7015f702e43d2eec6457926f29905b242def9b2f9e423340ce74bd3ecf8a1: Status 404 returned error can't find the container with id 30c7015f702e43d2eec6457926f29905b242def9b2f9e423340ce74bd3ecf8a1 Nov 22 04:49:57 crc kubenswrapper[4818]: W1122 04:49:57.830467 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4bfeb5de_dcfb_4f0b_b118_112ee3262c18.slice/crio-0df1f8a9fd97b1cadc3802e231d697eacbca53b212d72f6eb21673eb94ab06c4 WatchSource:0}: Error finding container 0df1f8a9fd97b1cadc3802e231d697eacbca53b212d72f6eb21673eb94ab06c4: Status 404 returned error can't find the container with id 0df1f8a9fd97b1cadc3802e231d697eacbca53b212d72f6eb21673eb94ab06c4 Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.847848 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-b86wj" Nov 22 04:49:57 crc kubenswrapper[4818]: I1122 04:49:57.903570 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:49:57 crc kubenswrapper[4818]: E1122 04:49:57.903930 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:49:58.403915243 +0000 UTC m=+150.978331760 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.004832 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:58 crc kubenswrapper[4818]: E1122 04:49:58.005473 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:49:58.505461431 +0000 UTC m=+151.079877958 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.028189 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr"] Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.042139 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb"] Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.048488 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nz99w"] Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.051118 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7bmhw"] Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.066054 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jtg8k"] Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.066625 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-jp5fs"] Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.095453 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dhhh4"] Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.098390 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fxwd8"] Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.103979 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rsc75"] Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.106103 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:49:58 crc kubenswrapper[4818]: E1122 04:49:58.106477 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:49:58.606462746 +0000 UTC m=+151.180879273 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:58 crc kubenswrapper[4818]: W1122 04:49:58.119749 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb025bc9d_ac7e_4a60_bcdc_a6a156eb2c9a.slice/crio-7c6e780c9948749f9f9c182a603c588204e79d466847a828fcec560c0ccfb9eb WatchSource:0}: Error finding container 7c6e780c9948749f9f9c182a603c588204e79d466847a828fcec560c0ccfb9eb: Status 404 returned error can't find the container with id 7c6e780c9948749f9f9c182a603c588204e79d466847a828fcec560c0ccfb9eb Nov 22 04:49:58 crc kubenswrapper[4818]: W1122 04:49:58.168276 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb6893d2_46c1_46fa_8c40_5bbd900d9fc1.slice/crio-7db6ec31496c67e0e7acd7957e2415c93ef8ecb42a307ba76b43e78517cea32d WatchSource:0}: Error finding container 7db6ec31496c67e0e7acd7957e2415c93ef8ecb42a307ba76b43e78517cea32d: Status 404 returned error can't find the container with id 7db6ec31496c67e0e7acd7957e2415c93ef8ecb42a307ba76b43e78517cea32d Nov 22 04:49:58 crc kubenswrapper[4818]: W1122 04:49:58.171895 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4aa421fd_173a_4c0d_8c66_ddcb7b173808.slice/crio-5c33040ad69eef9e15c50bdf2f3690ec17330e723749820caa979e9fd494fd93 WatchSource:0}: Error finding container 5c33040ad69eef9e15c50bdf2f3690ec17330e723749820caa979e9fd494fd93: Status 404 returned error can't find the container with id 5c33040ad69eef9e15c50bdf2f3690ec17330e723749820caa979e9fd494fd93 Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.210222 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5484670-b8e2-40b9-a36b-3fb47bb89421-serving-cert\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.210590 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.210623 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d3b850f-8917-46af-b713-891a7c979d2e-config\") pod \"machine-api-operator-5694c8668f-8kzxl\" (UID: \"8d3b850f-8917-46af-b713-891a7c979d2e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8kzxl" Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.210661 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-config\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.210750 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-client-ca\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:58 crc kubenswrapper[4818]: E1122 04:49:58.211782 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:49:58.711760927 +0000 UTC m=+151.286177454 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.212631 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-client-ca\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.214213 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d3b850f-8917-46af-b713-891a7c979d2e-config\") pod \"machine-api-operator-5694c8668f-8kzxl\" (UID: \"8d3b850f-8917-46af-b713-891a7c979d2e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8kzxl" Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.208294 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rr7gt" event={"ID":"215b4f1f-75ab-4c5c-aca8-26ea0edf5978","Type":"ContainerStarted","Data":"3b1f7339dcbbc8f8f5fedcba7d6e58f353044d40be32c64d97afdf0a4864aa7a"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.218088 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-config\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.218793 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-rcvlq" event={"ID":"61feab0c-5f50-49fe-bf35-de2a5f66bf32","Type":"ContainerStarted","Data":"4e47b7e752719f7b0e5755e3fa9f73a4fb266ff462c4ce5a897145e2e88c3f56"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.218826 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-rcvlq" event={"ID":"61feab0c-5f50-49fe-bf35-de2a5f66bf32","Type":"ContainerStarted","Data":"25f1c482cbc9a71e949591a4a7580ffc491eea695aeb1a53996d91d3ca647259"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.219334 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-rcvlq" Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.225685 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr" event={"ID":"3e2ce5bb-1abf-47c6-b6d2-59389118cff0","Type":"ContainerStarted","Data":"e29db2ef2a75ba70396362382ab93ef1cc1cc84c2aee1d2cab6caa9b635bbc00"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.227263 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jtg8k" event={"ID":"4aa421fd-173a-4c0d-8c66-ddcb7b173808","Type":"ContainerStarted","Data":"5c33040ad69eef9e15c50bdf2f3690ec17330e723749820caa979e9fd494fd93"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.230465 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.230515 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.252530 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5484670-b8e2-40b9-a36b-3fb47bb89421-serving-cert\") pod \"controller-manager-879f6c89f-tjh65\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.256591 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-5gfmz" event={"ID":"eb0cffc3-bf07-411c-822e-aa45c58387cb","Type":"ContainerStarted","Data":"fb1971509071ef1da52ad1c96f24b52d6379eae2a7409b5da1fd9ef1a0d10f1b"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.256638 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-5gfmz" event={"ID":"eb0cffc3-bf07-411c-822e-aa45c58387cb","Type":"ContainerStarted","Data":"3e9ec2e43efb3e7a078e0a9eb343da658addef895b2fb8c9b31bff90bcdba0d8"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.257619 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-b86wj"] Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.258907 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8"] Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.264758 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-z7mtd" event={"ID":"6f239350-c7bc-4fe4-9c24-ca0070a0775c","Type":"ContainerStarted","Data":"0a2972184e4d8907d9b63901e289d476c43a9efc8b1e1282c50bf1bec0ece7ab"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.264796 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-z7mtd" event={"ID":"6f239350-c7bc-4fe4-9c24-ca0070a0775c","Type":"ContainerStarted","Data":"c3fd056ee3617efe8dc96bc0a4b943722333c368da35a8894abd8c47e4bd8ad9"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.267329 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq"] Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.268045 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" event={"ID":"a33da900-6b7c-4157-8a6e-fde1cddf51ca","Type":"ContainerStarted","Data":"53f66f9664cbddeb493eb02058a450b1088d5ea4f37953e767556c5937bffbf3"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.269392 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb" event={"ID":"5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe","Type":"ContainerStarted","Data":"06f582d65e046db969af6d3b830161762ac3115a138c9142df6f95f82d277316"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.269833 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqnns"] Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.271292 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zftfz"] Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.272040 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rsc75" event={"ID":"f2a1ecdb-4740-43c4-ba73-19d1f15465e1","Type":"ContainerStarted","Data":"f8d3e654c6e8b84ad2590afc63c8b379eeb50f8399bf5901948f39a1a716b790"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.272708 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4"] Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.278343 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.279859 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-dvt5z" event={"ID":"4bfeb5de-dcfb-4f0b-b118-112ee3262c18","Type":"ContainerStarted","Data":"0df1f8a9fd97b1cadc3802e231d697eacbca53b212d72f6eb21673eb94ab06c4"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.281495 4818 generic.go:334] "Generic (PLEG): container finished" podID="10918dd9-6412-4172-a0f3-eac3fbd10da1" containerID="dd342cbe90750aa931f58b65122af0542e04ef2bb148f772f7206413a30883d1" exitCode=0 Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.281789 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" event={"ID":"10918dd9-6412-4172-a0f3-eac3fbd10da1","Type":"ContainerDied","Data":"dd342cbe90750aa931f58b65122af0542e04ef2bb148f772f7206413a30883d1"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.284125 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" event={"ID":"40771f13-f771-4e07-8ad9-668a3e48fb3b","Type":"ContainerStarted","Data":"d321eef763dec71021c5deef61e5203f35e7e5d4391a11f10407658f5ac04cbe"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.286218 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7bmhw" event={"ID":"7ab520a0-54e7-4923-867c-215fffacf270","Type":"ContainerStarted","Data":"32efdf7907684cacd71a3348b41bfb959e98b944d7301a78c7bb2c31c1e4d5aa"} Nov 22 04:49:58 crc kubenswrapper[4818]: W1122 04:49:58.297062 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32f920f7_5adb_4cf3_bc1f_3a461b79bb10.slice/crio-1192d88ed09ea6b7c992a985a0fb363eb8a25cc6b43c36c8bf11d0882bb68062 WatchSource:0}: Error finding container 1192d88ed09ea6b7c992a985a0fb363eb8a25cc6b43c36c8bf11d0882bb68062: Status 404 returned error can't find the container with id 1192d88ed09ea6b7c992a985a0fb363eb8a25cc6b43c36c8bf11d0882bb68062 Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.297868 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xvd5r" event={"ID":"e48b2f0e-a02f-48c6-b513-33537b0108b6","Type":"ContainerStarted","Data":"01e33bacf2fcad1454da432bac99ccc3d9e5e3467dae1a0e17c7ea81d216a96e"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.297897 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xvd5r" event={"ID":"e48b2f0e-a02f-48c6-b513-33537b0108b6","Type":"ContainerStarted","Data":"d7bf003163f3dfb47b200897cb2a5aa4e46c9a1ee263cf297c2fd425ef6484b8"} Nov 22 04:49:58 crc kubenswrapper[4818]: W1122 04:49:58.304854 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c429c72_1ae2_4b7d_9786_57ed9341bd48.slice/crio-9a50300868d90623ead62228eff1cbce1426862742e8e05a0a38bb5b69e1b69a WatchSource:0}: Error finding container 9a50300868d90623ead62228eff1cbce1426862742e8e05a0a38bb5b69e1b69a: Status 404 returned error can't find the container with id 9a50300868d90623ead62228eff1cbce1426862742e8e05a0a38bb5b69e1b69a Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.308901 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-8kzxl" Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.310529 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" event={"ID":"1d4ef859-1409-451a-a2e1-4c1c8c6d473b","Type":"ContainerStarted","Data":"609bf2f02463db58829b6c51aec6fc3185c533005fbe9a89964428fa4c396351"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.312690 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:49:58 crc kubenswrapper[4818]: E1122 04:49:58.313289 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:49:58.813270346 +0000 UTC m=+151.387686873 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.313448 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:58 crc kubenswrapper[4818]: E1122 04:49:58.313792 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:49:58.81378576 +0000 UTC m=+151.388202287 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:58 crc kubenswrapper[4818]: W1122 04:49:58.314284 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde7a08c9_b456_475e_a688_56edc28603ac.slice/crio-78e8c85a7c2b153dbb359fbcfdedb9b928a16d81af64f850716453d954086359 WatchSource:0}: Error finding container 78e8c85a7c2b153dbb359fbcfdedb9b928a16d81af64f850716453d954086359: Status 404 returned error can't find the container with id 78e8c85a7c2b153dbb359fbcfdedb9b928a16d81af64f850716453d954086359 Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.314954 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-jp5fs" event={"ID":"eb6893d2-46c1-46fa-8c40-5bbd900d9fc1","Type":"ContainerStarted","Data":"7db6ec31496c67e0e7acd7957e2415c93ef8ecb42a307ba76b43e78517cea32d"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.320571 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j7ttg" event={"ID":"62bdffb2-a1c7-4960-81aa-b570834f78cd","Type":"ContainerStarted","Data":"ed14a9a7afba06d66c0fab3da68bbf9702a0a77ecbbacc3da8f3a4d59fff1f54"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.323198 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" event={"ID":"353ed841-0d3b-46df-beec-1906f6d801fd","Type":"ContainerStarted","Data":"54cea90252e2fa553613928256095d4621431b359ca2c50a528b05479c915594"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.333383 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-s8xvt" event={"ID":"5a66393b-1460-4ef7-9c8a-2001a0d187fb","Type":"ContainerStarted","Data":"17281908eb5d56f093cb6137553b40204fd23eb52e0d2ea894b5e11bfad05710"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.333421 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-s8xvt" event={"ID":"5a66393b-1460-4ef7-9c8a-2001a0d187fb","Type":"ContainerStarted","Data":"9fce274b3a664567daa4924406bc0d10c0f35b5ed9b01cafb629f4ffe4ec7459"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.340489 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q6xbv" event={"ID":"d7c47e41-70c7-4784-aa37-c609650e3e1b","Type":"ContainerStarted","Data":"33852d4e0964ac6d1fd5c4b47a8f533290183254fbb545eeb28497973d5d16cd"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.340522 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q6xbv" event={"ID":"d7c47e41-70c7-4784-aa37-c609650e3e1b","Type":"ContainerStarted","Data":"83115fbefaf55613242ae7b5aac6af78bbdb5e1ac9c4f9e40aac3ada672f0b9d"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.342672 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" event={"ID":"0dcaa80a-8993-4452-9acb-b8010eade7c6","Type":"ContainerStarted","Data":"7ac0f02857258c046abc9f0e943dc57129e4b404eb59e0b7b31e34bc54c99a9c"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.342701 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" event={"ID":"0dcaa80a-8993-4452-9acb-b8010eade7c6","Type":"ContainerStarted","Data":"2e962e5be8a55f03f46497d8f19ecb27be2fa251de108aaddc747db3eaec4fc0"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.343558 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.345574 4818 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-mkgj4 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.15:6443/healthz\": dial tcp 10.217.0.15:6443: connect: connection refused" start-of-body= Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.345922 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" podUID="0dcaa80a-8993-4452-9acb-b8010eade7c6" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.15:6443/healthz\": dial tcp 10.217.0.15:6443: connect: connection refused" Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.349298 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-887w5" event={"ID":"45ba8e80-12bf-472e-a0b7-b25601a9074e","Type":"ContainerStarted","Data":"2f60050c21a00c2116c5cdc384b218e8e82be9e14cc3095c5d61dadd46d2f357"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.351767 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-h56w2" event={"ID":"670a3117-320e-477c-8c4b-a54fe5a4ae9c","Type":"ContainerStarted","Data":"41638750cf72db58a7884e847e359081b79c6ed5fa16f76b13d3e295dcb0fed8"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.351836 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-h56w2" event={"ID":"670a3117-320e-477c-8c4b-a54fe5a4ae9c","Type":"ContainerStarted","Data":"2e51a4453f0a1e191a3e8883bc112a3d3f59a8d79ec7c58212962a629882a994"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.354808 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ps29" event={"ID":"85928d04-1176-4b1d-8bb7-d55376576225","Type":"ContainerStarted","Data":"fc7e11e7b88a9c9147409bf5f148c02e5f194a56a1e68249c68dee8ef5134ea7"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.354838 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ps29" event={"ID":"85928d04-1176-4b1d-8bb7-d55376576225","Type":"ContainerStarted","Data":"98ffe37d56d4042398e2ae7594a86f215f0abd1fcbf6ca8b7e917f524dbd46bb"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.357639 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-ccd5x" event={"ID":"2fde1815-9a0b-4f85-820c-b482ccdeab3c","Type":"ContainerStarted","Data":"eeb75cbfed1ad06fcf4402f18789020edabcec43c5088f6972c64093e28d8e86"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.360329 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-44grv" event={"ID":"7337e50a-daf8-4f50-9060-ec2a14752444","Type":"ContainerStarted","Data":"9a7dd9a7b540ae07d618f7dd3eaaf4d6dc573f83b4ebad59ba2feb6d8a3e0d36"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.360354 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-44grv" event={"ID":"7337e50a-daf8-4f50-9060-ec2a14752444","Type":"ContainerStarted","Data":"b5a06c0515651fa6bbe3ea5217613691a8567c2d13ae24e0ff27ae583554ab5c"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.361422 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fxwd8" event={"ID":"b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a","Type":"ContainerStarted","Data":"7c6e780c9948749f9f9c182a603c588204e79d466847a828fcec560c0ccfb9eb"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.363202 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-28f5t" event={"ID":"312c389c-57c0-4709-9320-777e4659d499","Type":"ContainerStarted","Data":"298746de9d72372e698dc05c7c1fc8a5af1923592f1617414ff61995a759bcb7"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.363316 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-28f5t" event={"ID":"312c389c-57c0-4709-9320-777e4659d499","Type":"ContainerStarted","Data":"667390d1d7c99299c79b523eb14105e29c3129cc05a4008d004c4a023428e6a7"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.365710 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-9nsph" event={"ID":"091bb385-63c1-443f-9aec-bedd08514ee8","Type":"ContainerStarted","Data":"30c7015f702e43d2eec6457926f29905b242def9b2f9e423340ce74bd3ecf8a1"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.369124 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-4pxtx"] Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.372515 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h7srs" event={"ID":"968b0694-9a30-49ca-bee7-d1eadbb31505","Type":"ContainerStarted","Data":"ed7462cbdd1365a3da893d144822c8339fa0df31afa6d648beaa464aceb630ee"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.372546 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h7srs" event={"ID":"968b0694-9a30-49ca-bee7-d1eadbb31505","Type":"ContainerStarted","Data":"545279f6bfb64307789783ef02bc16ee3b3427ca670a1604d0ea9906cdcc7515"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.379573 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7f58c" event={"ID":"23111038-ec66-4295-a6eb-b847c52a0c81","Type":"ContainerStarted","Data":"5c8f8bb4a07bd9700e57114b0d6772f7fa2942ed7ada88e634f274c4f1c93f39"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.386034 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6" event={"ID":"1c05f6d6-e233-46ca-91df-eaba2614018f","Type":"ContainerStarted","Data":"d824d518d53d63b2b6af7acb34dd5c86e161138fa268dc89b6090b3b2f5b39c2"} Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.414323 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:49:58 crc kubenswrapper[4818]: E1122 04:49:58.414995 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:49:58.9149701 +0000 UTC m=+151.489386617 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.434111 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.437564 4818 patch_prober.go:28] interesting pod/router-default-5444994796-5gfmz container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.437598 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5gfmz" podUID="eb0cffc3-bf07-411c-822e-aa45c58387cb" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.516651 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:58 crc kubenswrapper[4818]: E1122 04:49:58.517264 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:49:59.017236199 +0000 UTC m=+151.591652726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.627011 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:49:58 crc kubenswrapper[4818]: E1122 04:49:58.627324 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:49:59.127309259 +0000 UTC m=+151.701725786 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.727948 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:58 crc kubenswrapper[4818]: E1122 04:49:58.728615 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:49:59.228603152 +0000 UTC m=+151.803019679 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.835811 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:49:58 crc kubenswrapper[4818]: E1122 04:49:58.836193 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:49:59.336173605 +0000 UTC m=+151.910590182 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:58 crc kubenswrapper[4818]: I1122 04:49:58.939082 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:58 crc kubenswrapper[4818]: E1122 04:49:58.939540 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:49:59.439526813 +0000 UTC m=+152.013943340 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.044197 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:49:59 crc kubenswrapper[4818]: E1122 04:49:59.044935 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:49:59.544917328 +0000 UTC m=+152.119333855 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.046774 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-8kzxl"] Nov 22 04:49:59 crc kubenswrapper[4818]: W1122 04:49:59.069834 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d3b850f_8917_46af_b713_891a7c979d2e.slice/crio-6cd7caafb8f8b30844927f43c537bfe65ffcc2c573adafa4393f4244a3eef260 WatchSource:0}: Error finding container 6cd7caafb8f8b30844927f43c537bfe65ffcc2c573adafa4393f4244a3eef260: Status 404 returned error can't find the container with id 6cd7caafb8f8b30844927f43c537bfe65ffcc2c573adafa4393f4244a3eef260 Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.098069 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tjh65"] Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.115396 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-z7mtd" podStartSLOduration=5.115380232 podStartE2EDuration="5.115380232s" podCreationTimestamp="2025-11-22 04:49:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:59.11529817 +0000 UTC m=+151.689714697" watchObservedRunningTime="2025-11-22 04:49:59.115380232 +0000 UTC m=+151.689796759" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.146240 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:59 crc kubenswrapper[4818]: E1122 04:49:59.146603 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:49:59.64658449 +0000 UTC m=+152.221001027 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.194060 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" podStartSLOduration=128.19404245 podStartE2EDuration="2m8.19404245s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:59.164457876 +0000 UTC m=+151.738874403" watchObservedRunningTime="2025-11-22 04:49:59.19404245 +0000 UTC m=+151.768458977" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.246817 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:49:59 crc kubenswrapper[4818]: E1122 04:49:59.246980 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:49:59.746961798 +0000 UTC m=+152.321378325 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.247140 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:59 crc kubenswrapper[4818]: E1122 04:49:59.253263 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:49:59.753217887 +0000 UTC m=+152.327634414 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.274802 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-28f5t" podStartSLOduration=127.274784934 podStartE2EDuration="2m7.274784934s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:59.272907252 +0000 UTC m=+151.847323779" watchObservedRunningTime="2025-11-22 04:49:59.274784934 +0000 UTC m=+151.849201461" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.316960 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-rcvlq" podStartSLOduration=128.31694622 podStartE2EDuration="2m8.31694622s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:59.315826699 +0000 UTC m=+151.890243226" watchObservedRunningTime="2025-11-22 04:49:59.31694622 +0000 UTC m=+151.891362747" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.348230 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:49:59 crc kubenswrapper[4818]: E1122 04:49:59.348645 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:49:59.84863 +0000 UTC m=+152.423046527 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.395558 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb" event={"ID":"5a43d852-9b9f-4bc5-a6ad-9dd1847b2fbe","Type":"ContainerStarted","Data":"428bfe58cfe6efe46e33293c8ea6634b787880b99e733275b8351a0ef1623c9c"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.396725 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr" event={"ID":"3e2ce5bb-1abf-47c6-b6d2-59389118cff0","Type":"ContainerStarted","Data":"c29ce8ad8a76185b90693d3c14e8b2c8046e4ecb03ce4e057747e212d2327ea5"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.396982 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.398468 4818 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-btgsr container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.398515 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr" podUID="3e2ce5bb-1abf-47c6-b6d2-59389118cff0" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.398965 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rsc75" event={"ID":"f2a1ecdb-4740-43c4-ba73-19d1f15465e1","Type":"ContainerStarted","Data":"ae70ffb071cce7d6d6dcddfabf16e54624b672a42c5914d4d059aecc682b9ae8"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.413475 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-dvt5z" event={"ID":"4bfeb5de-dcfb-4f0b-b118-112ee3262c18","Type":"ContainerStarted","Data":"24aeda218d95ec19febf31b11c6740744eefd45618e4917f87a38422e3a24273"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.419020 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-b86wj" event={"ID":"ffee77a3-faea-4547-989b-02aab97c5471","Type":"ContainerStarted","Data":"f9d3110b5035a9e1ce50dddee219d11a9801434431437dbd8126ac238c116bc2"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.431866 4818 generic.go:334] "Generic (PLEG): container finished" podID="091bb385-63c1-443f-9aec-bedd08514ee8" containerID="e7e1163509e7e2407e8602ee6fcb9b54b237cbc773e9bbf0c912b56eab128c46" exitCode=0 Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.432008 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-9nsph" event={"ID":"091bb385-63c1-443f-9aec-bedd08514ee8","Type":"ContainerDied","Data":"e7e1163509e7e2407e8602ee6fcb9b54b237cbc773e9bbf0c912b56eab128c46"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.436698 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-5gfmz" podStartSLOduration=127.436684943 podStartE2EDuration="2m7.436684943s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:59.4030772 +0000 UTC m=+151.977493727" watchObservedRunningTime="2025-11-22 04:49:59.436684943 +0000 UTC m=+152.011101470" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.438277 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xvd5r" podStartSLOduration=128.438271636 podStartE2EDuration="2m8.438271636s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:59.436115138 +0000 UTC m=+152.010531665" watchObservedRunningTime="2025-11-22 04:49:59.438271636 +0000 UTC m=+152.012688163" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.439208 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4pxtx" event={"ID":"54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea","Type":"ContainerStarted","Data":"4da7823ae2afcdcba5e6f86dc4c90f3c35d5117ba01bc706363d5f679203df46"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.439307 4818 patch_prober.go:28] interesting pod/router-default-5444994796-5gfmz container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.439336 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5gfmz" podUID="eb0cffc3-bf07-411c-822e-aa45c58387cb" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.440748 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-ccd5x" event={"ID":"2fde1815-9a0b-4f85-820c-b482ccdeab3c","Type":"ContainerStarted","Data":"3b0f0171b6c45d34cabce1227239c359b0862b3d24efe6edecb188b97ab221c0"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.444795 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-jp5fs" event={"ID":"eb6893d2-46c1-46fa-8c40-5bbd900d9fc1","Type":"ContainerStarted","Data":"6bf0dfe22295e3386116dbe61f0ec80fe4298d711ac0ea4abef42206345758d7"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.446173 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" event={"ID":"1d4ef859-1409-451a-a2e1-4c1c8c6d473b","Type":"ContainerStarted","Data":"1d9c4f4c749f569d569ffc735c3e3244d269e0b217dcbf2dafad509d813c3355"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.446751 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.448357 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" event={"ID":"353ed841-0d3b-46df-beec-1906f6d801fd","Type":"ContainerStarted","Data":"2c302fc90978cceb46d9b02614009b42c6092cfb0211756cb5b368a9f782e821"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.448808 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.448861 4818 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-nz99w container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.448882 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" podUID="1d4ef859-1409-451a-a2e1-4c1c8c6d473b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.449850 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.450004 4818 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-4lz4x container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.450022 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" podUID="353ed841-0d3b-46df-beec-1906f6d801fd" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 22 04:49:59 crc kubenswrapper[4818]: E1122 04:49:59.450189 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:49:59.95017938 +0000 UTC m=+152.524595907 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.451093 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq" event={"ID":"32f920f7-5adb-4cf3-bc1f-3a461b79bb10","Type":"ContainerStarted","Data":"1192d88ed09ea6b7c992a985a0fb363eb8a25cc6b43c36c8bf11d0882bb68062"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.451917 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4" event={"ID":"751d9838-4c34-4251-b3a6-1d808f98a034","Type":"ContainerStarted","Data":"0161b648d121799219d7a2b33f36d4578ab0b0aee754d072863e7e4ebbbbbbea"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.453682 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jtg8k" event={"ID":"4aa421fd-173a-4c0d-8c66-ddcb7b173808","Type":"ContainerStarted","Data":"fe543358b8c62830071250812c12373d9b6d22f526d7c4e8fc80ccacb8be317c"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.464783 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j7ttg" event={"ID":"62bdffb2-a1c7-4960-81aa-b570834f78cd","Type":"ContainerStarted","Data":"031beb86e580fc0c2fd235a74c6d821c9d988d86bcd8c34cabe16c39fee8869b"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.466498 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" event={"ID":"d5484670-b8e2-40b9-a36b-3fb47bb89421","Type":"ContainerStarted","Data":"fe11bd277dcae022d5866f9a2af875bb91519a95a25776307fc641ef2efa74f9"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.468410 4818 generic.go:334] "Generic (PLEG): container finished" podID="215b4f1f-75ab-4c5c-aca8-26ea0edf5978" containerID="9565f7a5db2a6d0251f340c11d8938cefee578654e3acc254370d0ae0b596ffb" exitCode=0 Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.468458 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rr7gt" event={"ID":"215b4f1f-75ab-4c5c-aca8-26ea0edf5978","Type":"ContainerDied","Data":"9565f7a5db2a6d0251f340c11d8938cefee578654e3acc254370d0ae0b596ffb"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.471425 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-887w5" event={"ID":"45ba8e80-12bf-472e-a0b7-b25601a9074e","Type":"ContainerStarted","Data":"cd5b39c085031b2fc56abc021b1952acee6e50ab059fc9bf91a246efeed86350"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.472581 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" event={"ID":"c9472804-f8b1-440a-b12e-d351630ed795","Type":"ContainerStarted","Data":"194ece256bd9f6c02c9e8c793b4d0b5002bee4a9ece91fc94777b2f3de5cb28f"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.473876 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6" event={"ID":"1c05f6d6-e233-46ca-91df-eaba2614018f","Type":"ContainerStarted","Data":"b5cab459ec5f57d062f8498121232660785ff2804827d2edb37928f459e122d8"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.475409 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqnns" event={"ID":"de7a08c9-b456-475e-a688-56edc28603ac","Type":"ContainerStarted","Data":"78e8c85a7c2b153dbb359fbcfdedb9b928a16d81af64f850716453d954086359"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.477073 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-h56w2" podStartSLOduration=128.47706238 podStartE2EDuration="2m8.47706238s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:59.476418423 +0000 UTC m=+152.050834980" watchObservedRunningTime="2025-11-22 04:49:59.47706238 +0000 UTC m=+152.051478907" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.479817 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7f58c" event={"ID":"23111038-ec66-4295-a6eb-b847c52a0c81","Type":"ContainerStarted","Data":"b16a1545d8c61b935cf960396242fd93e0fa189d0e93c716d5461bd081da2b74"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.481073 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zftfz" event={"ID":"7c429c72-1ae2-4b7d-9786-57ed9341bd48","Type":"ContainerStarted","Data":"9a50300868d90623ead62228eff1cbce1426862742e8e05a0a38bb5b69e1b69a"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.483161 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-8kzxl" event={"ID":"8d3b850f-8917-46af-b713-891a7c979d2e","Type":"ContainerStarted","Data":"6cd7caafb8f8b30844927f43c537bfe65ffcc2c573adafa4393f4244a3eef260"} Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.484504 4818 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-mkgj4 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.15:6443/healthz\": dial tcp 10.217.0.15:6443: connect: connection refused" start-of-body= Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.484562 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" podUID="0dcaa80a-8993-4452-9acb-b8010eade7c6" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.15:6443/healthz\": dial tcp 10.217.0.15:6443: connect: connection refused" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.485028 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.485060 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.485097 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-44grv" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.486847 4818 patch_prober.go:28] interesting pod/console-operator-58897d9998-44grv container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.486874 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-44grv" podUID="7337e50a-daf8-4f50-9060-ec2a14752444" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.511575 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h7srs" podStartSLOduration=128.511561168 podStartE2EDuration="2m8.511561168s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:59.50946291 +0000 UTC m=+152.083879437" watchObservedRunningTime="2025-11-22 04:49:59.511561168 +0000 UTC m=+152.085977695" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.550779 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:49:59 crc kubenswrapper[4818]: E1122 04:49:59.550919 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:00.050901627 +0000 UTC m=+152.625318154 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.551085 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:59 crc kubenswrapper[4818]: E1122 04:49:59.553406 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:00.053392845 +0000 UTC m=+152.627809372 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.557560 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-44grv" podStartSLOduration=128.557539858 podStartE2EDuration="2m8.557539858s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:59.556731135 +0000 UTC m=+152.131147672" watchObservedRunningTime="2025-11-22 04:49:59.557539858 +0000 UTC m=+152.131956405" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.652861 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:49:59 crc kubenswrapper[4818]: E1122 04:49:59.653042 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:00.153016502 +0000 UTC m=+152.727433029 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.653676 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.658738 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-887w5" podStartSLOduration=127.658725817 podStartE2EDuration="2m7.658725817s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:59.595892539 +0000 UTC m=+152.170309066" watchObservedRunningTime="2025-11-22 04:49:59.658725817 +0000 UTC m=+152.233142344" Nov 22 04:49:59 crc kubenswrapper[4818]: E1122 04:49:59.660426 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:00.160405732 +0000 UTC m=+152.734822259 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.662286 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ps29" podStartSLOduration=128.662270793 podStartE2EDuration="2m8.662270793s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:59.661502552 +0000 UTC m=+152.235919099" watchObservedRunningTime="2025-11-22 04:49:59.662270793 +0000 UTC m=+152.236687390" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.714881 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-s8xvt" podStartSLOduration=128.714865812 podStartE2EDuration="2m8.714865812s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:59.714189404 +0000 UTC m=+152.288605931" watchObservedRunningTime="2025-11-22 04:49:59.714865812 +0000 UTC m=+152.289282339" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.753421 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" podStartSLOduration=127.7534038 podStartE2EDuration="2m7.7534038s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:59.752726671 +0000 UTC m=+152.327143198" watchObservedRunningTime="2025-11-22 04:49:59.7534038 +0000 UTC m=+152.327820327" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.755408 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:49:59 crc kubenswrapper[4818]: E1122 04:49:59.755773 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:00.255757623 +0000 UTC m=+152.830174150 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.797673 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jtg8k" podStartSLOduration=127.797659782 podStartE2EDuration="2m7.797659782s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:59.795209885 +0000 UTC m=+152.369626412" watchObservedRunningTime="2025-11-22 04:49:59.797659782 +0000 UTC m=+152.372076309" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.835676 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-jp5fs" podStartSLOduration=127.835660054 podStartE2EDuration="2m7.835660054s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:59.833607919 +0000 UTC m=+152.408024456" watchObservedRunningTime="2025-11-22 04:49:59.835660054 +0000 UTC m=+152.410076591" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.857612 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:59 crc kubenswrapper[4818]: E1122 04:49:59.857984 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:00.357970091 +0000 UTC m=+152.932386628 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.875473 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" podStartSLOduration=127.875454996 podStartE2EDuration="2m7.875454996s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:59.871852198 +0000 UTC m=+152.446268725" watchObservedRunningTime="2025-11-22 04:49:59.875454996 +0000 UTC m=+152.449871523" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.916832 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr" podStartSLOduration=127.916800439 podStartE2EDuration="2m7.916800439s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:59.914091946 +0000 UTC m=+152.488508473" watchObservedRunningTime="2025-11-22 04:49:59.916800439 +0000 UTC m=+152.491216966" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.955808 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7f58c" podStartSLOduration=127.955790789 podStartE2EDuration="2m7.955790789s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:49:59.953115076 +0000 UTC m=+152.527531633" watchObservedRunningTime="2025-11-22 04:49:59.955790789 +0000 UTC m=+152.530207326" Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.958916 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:49:59 crc kubenswrapper[4818]: E1122 04:49:59.959068 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:00.459042957 +0000 UTC m=+153.033459484 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:49:59 crc kubenswrapper[4818]: I1122 04:49:59.959313 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:49:59 crc kubenswrapper[4818]: E1122 04:49:59.959613 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:00.459606543 +0000 UTC m=+153.034023070 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.039028 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rsc75" podStartSLOduration=129.039009921 podStartE2EDuration="2m9.039009921s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:00.036354669 +0000 UTC m=+152.610771216" watchObservedRunningTime="2025-11-22 04:50:00.039009921 +0000 UTC m=+152.613426458" Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.060983 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:00 crc kubenswrapper[4818]: E1122 04:50:00.061159 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:00.561133421 +0000 UTC m=+153.135549948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.061209 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:00 crc kubenswrapper[4818]: E1122 04:50:00.061534 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:00.561520753 +0000 UTC m=+153.135937290 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.162581 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:00 crc kubenswrapper[4818]: E1122 04:50:00.162734 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:00.662709502 +0000 UTC m=+153.237126029 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.162843 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:00 crc kubenswrapper[4818]: E1122 04:50:00.163125 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:00.663112733 +0000 UTC m=+153.237529260 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.263965 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:00 crc kubenswrapper[4818]: E1122 04:50:00.264142 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:00.764106227 +0000 UTC m=+153.338522784 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.264217 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:00 crc kubenswrapper[4818]: E1122 04:50:00.264654 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:00.764639882 +0000 UTC m=+153.339056449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.365701 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:00 crc kubenswrapper[4818]: E1122 04:50:00.366361 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:00.866346485 +0000 UTC m=+153.440763012 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.460922 4818 patch_prober.go:28] interesting pod/router-default-5444994796-5gfmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 04:50:00 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Nov 22 04:50:00 crc kubenswrapper[4818]: [+]process-running ok Nov 22 04:50:00 crc kubenswrapper[4818]: healthz check failed Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.461223 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5gfmz" podUID="eb0cffc3-bf07-411c-822e-aa45c58387cb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.467718 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:00 crc kubenswrapper[4818]: E1122 04:50:00.468060 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:00.968046499 +0000 UTC m=+153.542463026 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.491130 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zftfz" event={"ID":"7c429c72-1ae2-4b7d-9786-57ed9341bd48","Type":"ContainerStarted","Data":"c3ffdc6d322d23a0ed2202d20b3bc288651f01e77903019328c6692ad2e90957"} Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.492413 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-8kzxl" event={"ID":"8d3b850f-8917-46af-b713-891a7c979d2e","Type":"ContainerStarted","Data":"a22cd05cbe0f08ec5a6952117524ba218a7416d75615418661b1c47dad58cf06"} Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.493740 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q6xbv" event={"ID":"d7c47e41-70c7-4784-aa37-c609650e3e1b","Type":"ContainerStarted","Data":"ca2a14622520b2645caa3430239defa2510381d5277c4500e4b9064f00902b98"} Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.496037 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4pxtx" event={"ID":"54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea","Type":"ContainerStarted","Data":"9a4e3d9037a29c0eb2c149ef2e79933a6d72a5a243ce3a3d459da58590ba65a1"} Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.497472 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" event={"ID":"c9472804-f8b1-440a-b12e-d351630ed795","Type":"ContainerStarted","Data":"16cb33f5e88dcd79845d4edbdb7ae25e5e786c05ecaa84cb2f9edef7dc3e251c"} Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.498518 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq" event={"ID":"32f920f7-5adb-4cf3-bc1f-3a461b79bb10","Type":"ContainerStarted","Data":"71599d9bb135a7ece64fc1b089fb835e2d2f0d89cb02dc31b9d10eddd292e1e8"} Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.499684 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6" event={"ID":"1c05f6d6-e233-46ca-91df-eaba2614018f","Type":"ContainerStarted","Data":"024e0975355781d569be554d37755a3001e5097dbbd3daed1eee79ffe8c8ac03"} Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.501361 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j7ttg" event={"ID":"62bdffb2-a1c7-4960-81aa-b570834f78cd","Type":"ContainerStarted","Data":"ec9081cb238de9e7b81be9ac4b1a636fea6a97961ca924fce1f3592c24503d35"} Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.502516 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-dvt5z" event={"ID":"4bfeb5de-dcfb-4f0b-b118-112ee3262c18","Type":"ContainerStarted","Data":"ece49587c40fb94a992b08c7c052ae17e2a140559be08acaa88104c05b37d726"} Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.503941 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rr7gt" event={"ID":"215b4f1f-75ab-4c5c-aca8-26ea0edf5978","Type":"ContainerStarted","Data":"4b561bad8f7c14e953195759b22ef04dabed2c8b4a969afb9bfbc3289956eecf"} Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.505521 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" event={"ID":"10918dd9-6412-4172-a0f3-eac3fbd10da1","Type":"ContainerStarted","Data":"da8e24ae93a534055a483075fbffd080e97a90487d471ac4693073b3d10eed0c"} Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.508696 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-ccd5x" podStartSLOduration=6.508683303 podStartE2EDuration="6.508683303s" podCreationTimestamp="2025-11-22 04:49:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:00.113664649 +0000 UTC m=+152.688081176" watchObservedRunningTime="2025-11-22 04:50:00.508683303 +0000 UTC m=+153.083099840" Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.510846 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7bmhw" event={"ID":"7ab520a0-54e7-4923-867c-215fffacf270","Type":"ContainerStarted","Data":"ef6bf3266d5975132c13f52cbf1d58065406f7308a97ce3a11c163c6b1051227"} Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.510964 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q6xbv" podStartSLOduration=128.510954365 podStartE2EDuration="2m8.510954365s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:00.510689187 +0000 UTC m=+153.085105714" watchObservedRunningTime="2025-11-22 04:50:00.510954365 +0000 UTC m=+153.085370892" Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.512785 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fxwd8" event={"ID":"b025bc9d-ac7e-4a60-bcdc-a6a156eb2c9a","Type":"ContainerStarted","Data":"9b07208bfb3c1bb5f14fc1c3706744e1031fc3e9cd7996370b0a1d1765b06845"} Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.513766 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4" event={"ID":"751d9838-4c34-4251-b3a6-1d808f98a034","Type":"ContainerStarted","Data":"b7ccf6acf2f43368232355bb645f785a164ce14e18e91b3babade5df387c1565"} Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.516172 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" event={"ID":"d5484670-b8e2-40b9-a36b-3fb47bb89421","Type":"ContainerStarted","Data":"726de57de72645b656493b838692c8024ebb101a30d9def67c7ffc96630fd79c"} Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.517156 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" event={"ID":"40771f13-f771-4e07-8ad9-668a3e48fb3b","Type":"ContainerStarted","Data":"4093199f9007e024feb3fec34f0535b52d414851aeb9ef3460272a578e88a426"} Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.518099 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-b86wj" event={"ID":"ffee77a3-faea-4547-989b-02aab97c5471","Type":"ContainerStarted","Data":"2216698f96c10ec960d2a990fa4ba74e7891d486c97bfb6fda60d75272cfb712"} Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.521295 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-9nsph" event={"ID":"091bb385-63c1-443f-9aec-bedd08514ee8","Type":"ContainerStarted","Data":"3fe155474fcddb72df6b087198b163dc0a236e6f1f98f91050a6cecd58f476ea"} Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.522434 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqnns" event={"ID":"de7a08c9-b456-475e-a688-56edc28603ac","Type":"ContainerStarted","Data":"6029e33f2e86c237ba5fa2cd39c1c879a023a562411e2529029cd103590eb201"} Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.523093 4818 patch_prober.go:28] interesting pod/console-operator-58897d9998-44grv container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.523128 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-44grv" podUID="7337e50a-daf8-4f50-9060-ec2a14752444" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.523354 4818 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-nz99w container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.523393 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" podUID="1d4ef859-1409-451a-a2e1-4c1c8c6d473b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.523988 4818 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-mkgj4 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.15:6443/healthz\": dial tcp 10.217.0.15:6443: connect: connection refused" start-of-body= Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.524048 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" podUID="0dcaa80a-8993-4452-9acb-b8010eade7c6" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.15:6443/healthz\": dial tcp 10.217.0.15:6443: connect: connection refused" Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.523992 4818 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-4lz4x container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.524102 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" podUID="353ed841-0d3b-46df-beec-1906f6d801fd" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.524454 4818 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-btgsr container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.524526 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr" podUID="3e2ce5bb-1abf-47c6-b6d2-59389118cff0" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.532138 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-92bz6" podStartSLOduration=128.53211876 podStartE2EDuration="2m8.53211876s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:00.530163567 +0000 UTC m=+153.104580114" watchObservedRunningTime="2025-11-22 04:50:00.53211876 +0000 UTC m=+153.106535287" Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.564378 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb" podStartSLOduration=128.564361546 podStartE2EDuration="2m8.564361546s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:00.550837429 +0000 UTC m=+153.125253966" watchObservedRunningTime="2025-11-22 04:50:00.564361546 +0000 UTC m=+153.138778083" Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.565724 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4" podStartSLOduration=129.565716793 podStartE2EDuration="2m9.565716793s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:00.564241733 +0000 UTC m=+153.138658260" watchObservedRunningTime="2025-11-22 04:50:00.565716793 +0000 UTC m=+153.140133340" Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.568803 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:00 crc kubenswrapper[4818]: E1122 04:50:00.568963 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.068949421 +0000 UTC m=+153.643365938 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.569048 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:00 crc kubenswrapper[4818]: E1122 04:50:00.569317 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.06931082 +0000 UTC m=+153.643727347 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.670370 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:00 crc kubenswrapper[4818]: E1122 04:50:00.670550 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.170523831 +0000 UTC m=+153.744940348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.678062 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:00 crc kubenswrapper[4818]: E1122 04:50:00.678688 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.178670912 +0000 UTC m=+153.753087439 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.780435 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:00 crc kubenswrapper[4818]: E1122 04:50:00.780613 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.280590652 +0000 UTC m=+153.855007179 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.780880 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:00 crc kubenswrapper[4818]: E1122 04:50:00.781202 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.281187769 +0000 UTC m=+153.855604296 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.884374 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:00 crc kubenswrapper[4818]: E1122 04:50:00.884467 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.384451194 +0000 UTC m=+153.958867721 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.885042 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:00 crc kubenswrapper[4818]: E1122 04:50:00.892061 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.392041651 +0000 UTC m=+153.966458178 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.994894 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:00 crc kubenswrapper[4818]: E1122 04:50:00.995444 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.49542449 +0000 UTC m=+154.069841017 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:00 crc kubenswrapper[4818]: I1122 04:50:00.995538 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:00 crc kubenswrapper[4818]: E1122 04:50:00.995870 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.495848071 +0000 UTC m=+154.070264598 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.096499 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:01 crc kubenswrapper[4818]: E1122 04:50:01.096716 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.596686141 +0000 UTC m=+154.171102678 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.096887 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:01 crc kubenswrapper[4818]: E1122 04:50:01.097265 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.597235706 +0000 UTC m=+154.171652323 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.198007 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:01 crc kubenswrapper[4818]: E1122 04:50:01.198374 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.698359995 +0000 UTC m=+154.272776522 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.299109 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:01 crc kubenswrapper[4818]: E1122 04:50:01.299434 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.799419971 +0000 UTC m=+154.373836498 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.400409 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:01 crc kubenswrapper[4818]: E1122 04:50:01.400607 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.900576419 +0000 UTC m=+154.474992956 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.400668 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.400814 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.400959 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:01 crc kubenswrapper[4818]: E1122 04:50:01.401328 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:01.90131756 +0000 UTC m=+154.475734097 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.401963 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.406996 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.437409 4818 patch_prober.go:28] interesting pod/router-default-5444994796-5gfmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 04:50:01 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Nov 22 04:50:01 crc kubenswrapper[4818]: [+]process-running ok Nov 22 04:50:01 crc kubenswrapper[4818]: healthz check failed Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.437473 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5gfmz" podUID="eb0cffc3-bf07-411c-822e-aa45c58387cb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.501672 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.501837 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.501879 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:50:01 crc kubenswrapper[4818]: E1122 04:50:01.502650 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:02.002623722 +0000 UTC m=+154.577040249 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.507734 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.508190 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.529770 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq" event={"ID":"32f920f7-5adb-4cf3-bc1f-3a461b79bb10","Type":"ContainerStarted","Data":"bf93732b16b710f50b73348dabbec83dcc1031d27c85887947b1e8481551a588"} Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.531899 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-8kzxl" event={"ID":"8d3b850f-8917-46af-b713-891a7c979d2e","Type":"ContainerStarted","Data":"c1d1bacbc9a2ddcaf40b89d62b9c7311f198090cac1e942b07220909fb0732eb"} Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.534193 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-9nsph" event={"ID":"091bb385-63c1-443f-9aec-bedd08514ee8","Type":"ContainerStarted","Data":"4dd311570e16f842e5d16efc79a0b752ec15918603c433b65a800ffa853a7f84"} Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.535594 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqnns" event={"ID":"de7a08c9-b456-475e-a688-56edc28603ac","Type":"ContainerStarted","Data":"4f39e0d43c124e98fab17e1cbeec08f4402f35ffdae6921a5f1fbe858074aac9"} Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.535655 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqnns" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.536783 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" event={"ID":"a33da900-6b7c-4157-8a6e-fde1cddf51ca","Type":"ContainerStarted","Data":"0caa5c78c591b4c3372f64d4c3bb8c744d0b22e14a6732c4076e3f9a66df726c"} Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.538330 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4pxtx" event={"ID":"54c2deaa-b6aa-4384-ac9b-fe41eb0a5bea","Type":"ContainerStarted","Data":"6ab521ea20a0a11edf8269263df4e594356dee9412d610c61bad4194f4a6224f"} Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.540407 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7bmhw" event={"ID":"7ab520a0-54e7-4923-867c-215fffacf270","Type":"ContainerStarted","Data":"9267945c8b369058c88d3ea41f0369834e7f451552d5bfcbcf06c7d9009606d7"} Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.543101 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-b86wj" event={"ID":"ffee77a3-faea-4547-989b-02aab97c5471","Type":"ContainerStarted","Data":"5477a948cb876556f30bfdb7e00bb8ab02c370c092f3e03c7449d84e7eecfab8"} Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.543138 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-b86wj" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.546327 4818 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-tjh65 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.546370 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" podUID="d5484670-b8e2-40b9-a36b-3fb47bb89421" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.546870 4818 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-nz99w container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.547055 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.547137 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" podUID="1d4ef859-1409-451a-a2e1-4c1c8c6d473b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.547171 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rr7gt" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.547186 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.547462 4818 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-k27q8 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.547491 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" podUID="c9472804-f8b1-440a-b12e-d351630ed795" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.571283 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vrvxq" podStartSLOduration=129.571266218 podStartE2EDuration="2m9.571266218s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:01.570466945 +0000 UTC m=+154.144883472" watchObservedRunningTime="2025-11-22 04:50:01.571266218 +0000 UTC m=+154.145682745" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.604482 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:01 crc kubenswrapper[4818]: E1122 04:50:01.604897 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:02.104883631 +0000 UTC m=+154.679300158 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.609654 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4pxtx" podStartSLOduration=129.60963855 podStartE2EDuration="2m9.60963855s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:01.603444832 +0000 UTC m=+154.177861369" watchObservedRunningTime="2025-11-22 04:50:01.60963855 +0000 UTC m=+154.184055077" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.616175 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.632201 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.661339 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" podStartSLOduration=129.661320184 podStartE2EDuration="2m9.661320184s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:01.627735032 +0000 UTC m=+154.202151569" watchObservedRunningTime="2025-11-22 04:50:01.661320184 +0000 UTC m=+154.235736711" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.662681 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rr7gt" podStartSLOduration=130.662674831 podStartE2EDuration="2m10.662674831s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:01.66264754 +0000 UTC m=+154.237064067" watchObservedRunningTime="2025-11-22 04:50:01.662674831 +0000 UTC m=+154.237091358" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.708752 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:01 crc kubenswrapper[4818]: E1122 04:50:01.708872 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:02.208846546 +0000 UTC m=+154.783263063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.709207 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:01 crc kubenswrapper[4818]: E1122 04:50:01.709457 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:02.209447393 +0000 UTC m=+154.783863920 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.723023 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.727824 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.742059 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-9nsph" podStartSLOduration=130.742042248 podStartE2EDuration="2m10.742042248s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:01.708208929 +0000 UTC m=+154.282625456" watchObservedRunningTime="2025-11-22 04:50:01.742042248 +0000 UTC m=+154.316458775" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.744022 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-b86wj" podStartSLOduration=7.744017561 podStartE2EDuration="7.744017561s" podCreationTimestamp="2025-11-22 04:49:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:01.742006567 +0000 UTC m=+154.316423094" watchObservedRunningTime="2025-11-22 04:50:01.744017561 +0000 UTC m=+154.318434088" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.810261 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-dvt5z" podStartSLOduration=129.810227401 podStartE2EDuration="2m9.810227401s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:01.773330678 +0000 UTC m=+154.347747205" watchObservedRunningTime="2025-11-22 04:50:01.810227401 +0000 UTC m=+154.384643928" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.811362 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:01 crc kubenswrapper[4818]: E1122 04:50:01.811939 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:02.311919707 +0000 UTC m=+154.886336234 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.868426 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.868481 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.869856 4818 patch_prober.go:28] interesting pod/apiserver-76f77b778f-9nsph container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.869924 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-9nsph" podUID="091bb385-63c1-443f-9aec-bedd08514ee8" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.883952 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zftfz" podStartSLOduration=129.883934134 podStartE2EDuration="2m9.883934134s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:01.808875685 +0000 UTC m=+154.383292212" watchObservedRunningTime="2025-11-22 04:50:01.883934134 +0000 UTC m=+154.458350651" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.914017 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:01 crc kubenswrapper[4818]: E1122 04:50:01.914321 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:02.414303289 +0000 UTC m=+154.988719816 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.955954 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-8kzxl" podStartSLOduration=129.955940861 podStartE2EDuration="2m9.955940861s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:01.892492736 +0000 UTC m=+154.466909263" watchObservedRunningTime="2025-11-22 04:50:01.955940861 +0000 UTC m=+154.530357388" Nov 22 04:50:01 crc kubenswrapper[4818]: I1122 04:50:01.956665 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j7ttg" podStartSLOduration=130.95666137 podStartE2EDuration="2m10.95666137s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:01.954354767 +0000 UTC m=+154.528771294" watchObservedRunningTime="2025-11-22 04:50:01.95666137 +0000 UTC m=+154.531077897" Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.015052 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:02 crc kubenswrapper[4818]: E1122 04:50:02.015165 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:02.515149299 +0000 UTC m=+155.089565826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.015414 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:02 crc kubenswrapper[4818]: E1122 04:50:02.015707 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:02.515700014 +0000 UTC m=+155.090116541 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.084694 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqnns" podStartSLOduration=130.084677889 podStartE2EDuration="2m10.084677889s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:02.037590589 +0000 UTC m=+154.612007116" watchObservedRunningTime="2025-11-22 04:50:02.084677889 +0000 UTC m=+154.659094416" Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.106366 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" podStartSLOduration=130.106345378 podStartE2EDuration="2m10.106345378s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:02.086602631 +0000 UTC m=+154.661019158" watchObservedRunningTime="2025-11-22 04:50:02.106345378 +0000 UTC m=+154.680761915" Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.116751 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:02 crc kubenswrapper[4818]: E1122 04:50:02.117094 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:02.617079529 +0000 UTC m=+155.191496056 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.131373 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-7bmhw" podStartSLOduration=130.131353577 podStartE2EDuration="2m10.131353577s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:02.107492049 +0000 UTC m=+154.681908576" watchObservedRunningTime="2025-11-22 04:50:02.131353577 +0000 UTC m=+154.705770094" Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.164267 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fxwd8" podStartSLOduration=130.16423938 podStartE2EDuration="2m10.16423938s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:02.130458403 +0000 UTC m=+154.704874930" watchObservedRunningTime="2025-11-22 04:50:02.16423938 +0000 UTC m=+154.738655907" Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.164515 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-dhhh4" podStartSLOduration=130.164512659 podStartE2EDuration="2m10.164512659s" podCreationTimestamp="2025-11-22 04:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:02.163199213 +0000 UTC m=+154.737615740" watchObservedRunningTime="2025-11-22 04:50:02.164512659 +0000 UTC m=+154.738929186" Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.220783 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" podStartSLOduration=131.220768297 podStartE2EDuration="2m11.220768297s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:02.197378351 +0000 UTC m=+154.771794878" watchObservedRunningTime="2025-11-22 04:50:02.220768297 +0000 UTC m=+154.795184824" Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.221944 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:02 crc kubenswrapper[4818]: E1122 04:50:02.222217 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:02.722205736 +0000 UTC m=+155.296622263 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.323644 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:02 crc kubenswrapper[4818]: E1122 04:50:02.324052 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:02.824035443 +0000 UTC m=+155.398451970 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.425183 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:02 crc kubenswrapper[4818]: E1122 04:50:02.425672 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:02.925657894 +0000 UTC m=+155.500074411 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.447577 4818 patch_prober.go:28] interesting pod/router-default-5444994796-5gfmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 04:50:02 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Nov 22 04:50:02 crc kubenswrapper[4818]: [+]process-running ok Nov 22 04:50:02 crc kubenswrapper[4818]: healthz check failed Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.447627 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5gfmz" podUID="eb0cffc3-bf07-411c-822e-aa45c58387cb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.529816 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:02 crc kubenswrapper[4818]: E1122 04:50:02.530121 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:03.030106973 +0000 UTC m=+155.604523500 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.577047 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"34c6819e96f1c6c06e98e0d75ccb77d8fbfeb86a34d62cf45389b6df35166a5d"} Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.586390 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"6051a200925d9b703234ff3342fce80f27146b5683358ffddeff7b7f14537762"} Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.588806 4818 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-tjh65 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.588848 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" podUID="d5484670-b8e2-40b9-a36b-3fb47bb89421" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.630893 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:02 crc kubenswrapper[4818]: E1122 04:50:02.631394 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:03.131382545 +0000 UTC m=+155.705799062 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.732050 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:02 crc kubenswrapper[4818]: E1122 04:50:02.733202 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:03.233180881 +0000 UTC m=+155.807597408 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:02 crc kubenswrapper[4818]: E1122 04:50:02.834517 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:03.334504424 +0000 UTC m=+155.908920951 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.834660 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:02 crc kubenswrapper[4818]: I1122 04:50:02.935693 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:02 crc kubenswrapper[4818]: E1122 04:50:02.935890 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:03.435874769 +0000 UTC m=+156.010291296 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.037645 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:03 crc kubenswrapper[4818]: E1122 04:50:03.038053 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:03.538036534 +0000 UTC m=+156.112453061 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.054607 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rr7gt" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.138885 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:03 crc kubenswrapper[4818]: E1122 04:50:03.139277 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:03.639262786 +0000 UTC m=+156.213679303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.241080 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:03 crc kubenswrapper[4818]: E1122 04:50:03.241369 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:03.74135765 +0000 UTC m=+156.315774177 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.311043 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.311629 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.324122 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.324294 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.342520 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.342815 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:03 crc kubenswrapper[4818]: E1122 04:50:03.342913 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:03.842899999 +0000 UTC m=+156.417316526 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.343111 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:03 crc kubenswrapper[4818]: E1122 04:50:03.343362 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:03.843355732 +0000 UTC m=+156.417772259 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.437795 4818 patch_prober.go:28] interesting pod/router-default-5444994796-5gfmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 04:50:03 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Nov 22 04:50:03 crc kubenswrapper[4818]: [+]process-running ok Nov 22 04:50:03 crc kubenswrapper[4818]: healthz check failed Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.437844 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5gfmz" podUID="eb0cffc3-bf07-411c-822e-aa45c58387cb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.444775 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.444984 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5bf364b3-3f72-4ccb-8286-26f5eec4e399-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5bf364b3-3f72-4ccb-8286-26f5eec4e399\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.445020 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5bf364b3-3f72-4ccb-8286-26f5eec4e399-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5bf364b3-3f72-4ccb-8286-26f5eec4e399\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 04:50:03 crc kubenswrapper[4818]: E1122 04:50:03.445101 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:03.945089756 +0000 UTC m=+156.519506283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.546043 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5bf364b3-3f72-4ccb-8286-26f5eec4e399-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5bf364b3-3f72-4ccb-8286-26f5eec4e399\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.546092 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5bf364b3-3f72-4ccb-8286-26f5eec4e399-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5bf364b3-3f72-4ccb-8286-26f5eec4e399\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.546135 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:03 crc kubenswrapper[4818]: E1122 04:50:03.546401 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:04.046391239 +0000 UTC m=+156.620807766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.546535 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5bf364b3-3f72-4ccb-8286-26f5eec4e399-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5bf364b3-3f72-4ccb-8286-26f5eec4e399\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.564953 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5bf364b3-3f72-4ccb-8286-26f5eec4e399-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5bf364b3-3f72-4ccb-8286-26f5eec4e399\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.587650 4818 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-k27q8 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.587695 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" podUID="c9472804-f8b1-440a-b12e-d351630ed795" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.597413 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"04849e6d5703eef350eddc2e6bcf63922f463b39db9778d31ba2c740b6fe900b"} Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.601319 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"92289f18f581055a32324670f2f8ce20917523161f603465ecde9247f3d28dfd"} Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.606113 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"8fa4e3383a9fc2ed26f7fa316b6e25c6a25f85a9997bf9423a76fb9a3c89e1c2"} Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.606135 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"56b581cbb03e6c418956ad984645147bc0b03a2fe6a92bbfde7655c5b9f0b9a4"} Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.606374 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.611617 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" event={"ID":"a33da900-6b7c-4157-8a6e-fde1cddf51ca","Type":"ContainerStarted","Data":"a7e2190fda37dac698b8f3e48051572e267b036e84c93d84ebd9df3c985342f3"} Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.638098 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.647626 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:03 crc kubenswrapper[4818]: E1122 04:50:03.647783 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:04.147760283 +0000 UTC m=+156.722176810 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.648001 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:03 crc kubenswrapper[4818]: E1122 04:50:03.648304 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:04.148297168 +0000 UTC m=+156.722713695 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.734377 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pg659"] Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.735543 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pg659" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.737672 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.748557 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.748894 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pg659"] Nov 22 04:50:03 crc kubenswrapper[4818]: E1122 04:50:03.749701 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:04.249682383 +0000 UTC m=+156.824098910 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.750808 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:03 crc kubenswrapper[4818]: E1122 04:50:03.751021 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:04.251014009 +0000 UTC m=+156.825430536 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.851838 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.852100 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30e236a6-d269-4f06-ae6e-e3c9bb74b4bf-utilities\") pod \"community-operators-pg659\" (UID: \"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf\") " pod="openshift-marketplace/community-operators-pg659" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.852144 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stfxm\" (UniqueName: \"kubernetes.io/projected/30e236a6-d269-4f06-ae6e-e3c9bb74b4bf-kube-api-access-stfxm\") pod \"community-operators-pg659\" (UID: \"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf\") " pod="openshift-marketplace/community-operators-pg659" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.852167 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30e236a6-d269-4f06-ae6e-e3c9bb74b4bf-catalog-content\") pod \"community-operators-pg659\" (UID: \"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf\") " pod="openshift-marketplace/community-operators-pg659" Nov 22 04:50:03 crc kubenswrapper[4818]: E1122 04:50:03.852302 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:04.352278211 +0000 UTC m=+156.926694738 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.918235 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-blsms"] Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.919455 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-blsms" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.925478 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.936921 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-blsms"] Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.955036 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30e236a6-d269-4f06-ae6e-e3c9bb74b4bf-utilities\") pod \"community-operators-pg659\" (UID: \"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf\") " pod="openshift-marketplace/community-operators-pg659" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.955121 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stfxm\" (UniqueName: \"kubernetes.io/projected/30e236a6-d269-4f06-ae6e-e3c9bb74b4bf-kube-api-access-stfxm\") pod \"community-operators-pg659\" (UID: \"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf\") " pod="openshift-marketplace/community-operators-pg659" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.955145 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30e236a6-d269-4f06-ae6e-e3c9bb74b4bf-catalog-content\") pod \"community-operators-pg659\" (UID: \"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf\") " pod="openshift-marketplace/community-operators-pg659" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.955207 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:03 crc kubenswrapper[4818]: E1122 04:50:03.955530 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:04.455519526 +0000 UTC m=+157.029936053 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.956362 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30e236a6-d269-4f06-ae6e-e3c9bb74b4bf-utilities\") pod \"community-operators-pg659\" (UID: \"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf\") " pod="openshift-marketplace/community-operators-pg659" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.956637 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30e236a6-d269-4f06-ae6e-e3c9bb74b4bf-catalog-content\") pod \"community-operators-pg659\" (UID: \"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf\") " pod="openshift-marketplace/community-operators-pg659" Nov 22 04:50:03 crc kubenswrapper[4818]: I1122 04:50:03.991185 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stfxm\" (UniqueName: \"kubernetes.io/projected/30e236a6-d269-4f06-ae6e-e3c9bb74b4bf-kube-api-access-stfxm\") pod \"community-operators-pg659\" (UID: \"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf\") " pod="openshift-marketplace/community-operators-pg659" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.051105 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pg659" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.058676 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.058861 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bclj5\" (UniqueName: \"kubernetes.io/projected/0e0614cc-57b5-4782-b462-e693c03c2117-kube-api-access-bclj5\") pod \"certified-operators-blsms\" (UID: \"0e0614cc-57b5-4782-b462-e693c03c2117\") " pod="openshift-marketplace/certified-operators-blsms" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.058903 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e0614cc-57b5-4782-b462-e693c03c2117-utilities\") pod \"certified-operators-blsms\" (UID: \"0e0614cc-57b5-4782-b462-e693c03c2117\") " pod="openshift-marketplace/certified-operators-blsms" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.058958 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e0614cc-57b5-4782-b462-e693c03c2117-catalog-content\") pod \"certified-operators-blsms\" (UID: \"0e0614cc-57b5-4782-b462-e693c03c2117\") " pod="openshift-marketplace/certified-operators-blsms" Nov 22 04:50:04 crc kubenswrapper[4818]: E1122 04:50:04.059131 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:04.559118102 +0000 UTC m=+157.133534629 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.119978 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.135241 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xsv9c"] Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.139911 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xsv9c" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.162068 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bclj5\" (UniqueName: \"kubernetes.io/projected/0e0614cc-57b5-4782-b462-e693c03c2117-kube-api-access-bclj5\") pod \"certified-operators-blsms\" (UID: \"0e0614cc-57b5-4782-b462-e693c03c2117\") " pod="openshift-marketplace/certified-operators-blsms" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.162120 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e0614cc-57b5-4782-b462-e693c03c2117-utilities\") pod \"certified-operators-blsms\" (UID: \"0e0614cc-57b5-4782-b462-e693c03c2117\") " pod="openshift-marketplace/certified-operators-blsms" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.162155 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.162190 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e0614cc-57b5-4782-b462-e693c03c2117-catalog-content\") pod \"certified-operators-blsms\" (UID: \"0e0614cc-57b5-4782-b462-e693c03c2117\") " pod="openshift-marketplace/certified-operators-blsms" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.162633 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e0614cc-57b5-4782-b462-e693c03c2117-catalog-content\") pod \"certified-operators-blsms\" (UID: \"0e0614cc-57b5-4782-b462-e693c03c2117\") " pod="openshift-marketplace/certified-operators-blsms" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.163134 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e0614cc-57b5-4782-b462-e693c03c2117-utilities\") pod \"certified-operators-blsms\" (UID: \"0e0614cc-57b5-4782-b462-e693c03c2117\") " pod="openshift-marketplace/certified-operators-blsms" Nov 22 04:50:04 crc kubenswrapper[4818]: E1122 04:50:04.163422 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:04.663410255 +0000 UTC m=+157.237826772 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.207581 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bclj5\" (UniqueName: \"kubernetes.io/projected/0e0614cc-57b5-4782-b462-e693c03c2117-kube-api-access-bclj5\") pod \"certified-operators-blsms\" (UID: \"0e0614cc-57b5-4782-b462-e693c03c2117\") " pod="openshift-marketplace/certified-operators-blsms" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.209332 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xsv9c"] Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.237141 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-blsms" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.262919 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.263142 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5d474c0-a83d-4383-a4d9-6fd0e128d34b-utilities\") pod \"community-operators-xsv9c\" (UID: \"b5d474c0-a83d-4383-a4d9-6fd0e128d34b\") " pod="openshift-marketplace/community-operators-xsv9c" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.263193 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5d474c0-a83d-4383-a4d9-6fd0e128d34b-catalog-content\") pod \"community-operators-xsv9c\" (UID: \"b5d474c0-a83d-4383-a4d9-6fd0e128d34b\") " pod="openshift-marketplace/community-operators-xsv9c" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.263239 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lggqt\" (UniqueName: \"kubernetes.io/projected/b5d474c0-a83d-4383-a4d9-6fd0e128d34b-kube-api-access-lggqt\") pod \"community-operators-xsv9c\" (UID: \"b5d474c0-a83d-4383-a4d9-6fd0e128d34b\") " pod="openshift-marketplace/community-operators-xsv9c" Nov 22 04:50:04 crc kubenswrapper[4818]: E1122 04:50:04.263377 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:04.763361772 +0000 UTC m=+157.337778299 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.363309 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p7spb"] Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.364169 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p7spb"] Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.364239 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p7spb" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.365135 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.365160 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5d474c0-a83d-4383-a4d9-6fd0e128d34b-utilities\") pod \"community-operators-xsv9c\" (UID: \"b5d474c0-a83d-4383-a4d9-6fd0e128d34b\") " pod="openshift-marketplace/community-operators-xsv9c" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.365200 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5d474c0-a83d-4383-a4d9-6fd0e128d34b-catalog-content\") pod \"community-operators-xsv9c\" (UID: \"b5d474c0-a83d-4383-a4d9-6fd0e128d34b\") " pod="openshift-marketplace/community-operators-xsv9c" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.365220 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lggqt\" (UniqueName: \"kubernetes.io/projected/b5d474c0-a83d-4383-a4d9-6fd0e128d34b-kube-api-access-lggqt\") pod \"community-operators-xsv9c\" (UID: \"b5d474c0-a83d-4383-a4d9-6fd0e128d34b\") " pod="openshift-marketplace/community-operators-xsv9c" Nov 22 04:50:04 crc kubenswrapper[4818]: E1122 04:50:04.365614 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:04.86560345 +0000 UTC m=+157.440019977 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.366355 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5d474c0-a83d-4383-a4d9-6fd0e128d34b-utilities\") pod \"community-operators-xsv9c\" (UID: \"b5d474c0-a83d-4383-a4d9-6fd0e128d34b\") " pod="openshift-marketplace/community-operators-xsv9c" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.366597 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5d474c0-a83d-4383-a4d9-6fd0e128d34b-catalog-content\") pod \"community-operators-xsv9c\" (UID: \"b5d474c0-a83d-4383-a4d9-6fd0e128d34b\") " pod="openshift-marketplace/community-operators-xsv9c" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.400213 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lggqt\" (UniqueName: \"kubernetes.io/projected/b5d474c0-a83d-4383-a4d9-6fd0e128d34b-kube-api-access-lggqt\") pod \"community-operators-xsv9c\" (UID: \"b5d474c0-a83d-4383-a4d9-6fd0e128d34b\") " pod="openshift-marketplace/community-operators-xsv9c" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.438085 4818 patch_prober.go:28] interesting pod/router-default-5444994796-5gfmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 04:50:04 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Nov 22 04:50:04 crc kubenswrapper[4818]: [+]process-running ok Nov 22 04:50:04 crc kubenswrapper[4818]: healthz check failed Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.438152 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5gfmz" podUID="eb0cffc3-bf07-411c-822e-aa45c58387cb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.468679 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.468934 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1146dedf-aeff-4812-84e5-dcfe1cfc56ca-catalog-content\") pod \"certified-operators-p7spb\" (UID: \"1146dedf-aeff-4812-84e5-dcfe1cfc56ca\") " pod="openshift-marketplace/certified-operators-p7spb" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.468960 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkf9w\" (UniqueName: \"kubernetes.io/projected/1146dedf-aeff-4812-84e5-dcfe1cfc56ca-kube-api-access-kkf9w\") pod \"certified-operators-p7spb\" (UID: \"1146dedf-aeff-4812-84e5-dcfe1cfc56ca\") " pod="openshift-marketplace/certified-operators-p7spb" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.469013 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1146dedf-aeff-4812-84e5-dcfe1cfc56ca-utilities\") pod \"certified-operators-p7spb\" (UID: \"1146dedf-aeff-4812-84e5-dcfe1cfc56ca\") " pod="openshift-marketplace/certified-operators-p7spb" Nov 22 04:50:04 crc kubenswrapper[4818]: E1122 04:50:04.469119 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:04.969101323 +0000 UTC m=+157.543517850 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.482505 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xsv9c" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.575190 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1146dedf-aeff-4812-84e5-dcfe1cfc56ca-utilities\") pod \"certified-operators-p7spb\" (UID: \"1146dedf-aeff-4812-84e5-dcfe1cfc56ca\") " pod="openshift-marketplace/certified-operators-p7spb" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.575241 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.575282 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1146dedf-aeff-4812-84e5-dcfe1cfc56ca-catalog-content\") pod \"certified-operators-p7spb\" (UID: \"1146dedf-aeff-4812-84e5-dcfe1cfc56ca\") " pod="openshift-marketplace/certified-operators-p7spb" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.575303 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkf9w\" (UniqueName: \"kubernetes.io/projected/1146dedf-aeff-4812-84e5-dcfe1cfc56ca-kube-api-access-kkf9w\") pod \"certified-operators-p7spb\" (UID: \"1146dedf-aeff-4812-84e5-dcfe1cfc56ca\") " pod="openshift-marketplace/certified-operators-p7spb" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.576216 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1146dedf-aeff-4812-84e5-dcfe1cfc56ca-utilities\") pod \"certified-operators-p7spb\" (UID: \"1146dedf-aeff-4812-84e5-dcfe1cfc56ca\") " pod="openshift-marketplace/certified-operators-p7spb" Nov 22 04:50:04 crc kubenswrapper[4818]: E1122 04:50:04.576512 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:05.07650206 +0000 UTC m=+157.650918587 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.576885 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1146dedf-aeff-4812-84e5-dcfe1cfc56ca-catalog-content\") pod \"certified-operators-p7spb\" (UID: \"1146dedf-aeff-4812-84e5-dcfe1cfc56ca\") " pod="openshift-marketplace/certified-operators-p7spb" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.612370 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkf9w\" (UniqueName: \"kubernetes.io/projected/1146dedf-aeff-4812-84e5-dcfe1cfc56ca-kube-api-access-kkf9w\") pod \"certified-operators-p7spb\" (UID: \"1146dedf-aeff-4812-84e5-dcfe1cfc56ca\") " pod="openshift-marketplace/certified-operators-p7spb" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.620044 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pg659"] Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.667776 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" event={"ID":"a33da900-6b7c-4157-8a6e-fde1cddf51ca","Type":"ContainerStarted","Data":"630109cacf7be1477afe1f905d46817b0bb7cfc6ebd2f1a7cceadb9dab0b2a3d"} Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.668028 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" event={"ID":"a33da900-6b7c-4157-8a6e-fde1cddf51ca","Type":"ContainerStarted","Data":"53528be0f8a9c6b299cc3ac9f32b5a44932ed764c3b6f62d925b9fada9eb773d"} Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.670365 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5bf364b3-3f72-4ccb-8286-26f5eec4e399","Type":"ContainerStarted","Data":"eedef202131044ba5571ada38de4b57707aa2d68bfae4fa5ae4d696b2edfae62"} Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.673008 4818 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.676729 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:04 crc kubenswrapper[4818]: E1122 04:50:04.676956 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:05.1769264 +0000 UTC m=+157.751342927 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.677086 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:04 crc kubenswrapper[4818]: E1122 04:50:04.677555 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:05.177543596 +0000 UTC m=+157.751960123 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.684775 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p7spb" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.725344 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-8zp8h" podStartSLOduration=10.725325875 podStartE2EDuration="10.725325875s" podCreationTimestamp="2025-11-22 04:49:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:04.723456944 +0000 UTC m=+157.297873471" watchObservedRunningTime="2025-11-22 04:50:04.725325875 +0000 UTC m=+157.299742402" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.778094 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:04 crc kubenswrapper[4818]: E1122 04:50:04.778815 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:05.278800828 +0000 UTC m=+157.853217355 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.819469 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.820408 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.828086 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.828343 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.848968 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.858173 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-blsms"] Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.888209 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:04 crc kubenswrapper[4818]: E1122 04:50:04.888486 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:05.388475828 +0000 UTC m=+157.962892355 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:04 crc kubenswrapper[4818]: W1122 04:50:04.905573 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e0614cc_57b5_4782_b462_e693c03c2117.slice/crio-1568c4aee9967dab35f69933192e224daed6ade8e9de040b5fa5a275d1664f09 WatchSource:0}: Error finding container 1568c4aee9967dab35f69933192e224daed6ade8e9de040b5fa5a275d1664f09: Status 404 returned error can't find the container with id 1568c4aee9967dab35f69933192e224daed6ade8e9de040b5fa5a275d1664f09 Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.989953 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.990262 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b4aabc4-c336-4114-86f4-4f94bb336609-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9b4aabc4-c336-4114-86f4-4f94bb336609\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 04:50:04 crc kubenswrapper[4818]: I1122 04:50:04.990323 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9b4aabc4-c336-4114-86f4-4f94bb336609-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9b4aabc4-c336-4114-86f4-4f94bb336609\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 04:50:04 crc kubenswrapper[4818]: E1122 04:50:04.990415 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:05.490401588 +0000 UTC m=+158.064818115 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.058994 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p7spb"] Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.091201 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b4aabc4-c336-4114-86f4-4f94bb336609-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9b4aabc4-c336-4114-86f4-4f94bb336609\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.091539 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.091571 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9b4aabc4-c336-4114-86f4-4f94bb336609-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9b4aabc4-c336-4114-86f4-4f94bb336609\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.091618 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9b4aabc4-c336-4114-86f4-4f94bb336609-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9b4aabc4-c336-4114-86f4-4f94bb336609\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 04:50:05 crc kubenswrapper[4818]: E1122 04:50:05.091776 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:05.591766202 +0000 UTC m=+158.166182729 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.095468 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.117707 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xsv9c"] Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.129234 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b4aabc4-c336-4114-86f4-4f94bb336609-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9b4aabc4-c336-4114-86f4-4f94bb336609\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.157214 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.195781 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:05 crc kubenswrapper[4818]: E1122 04:50:05.196372 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:05.696352224 +0000 UTC m=+158.270768761 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.297082 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:05 crc kubenswrapper[4818]: E1122 04:50:05.297738 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:05.797721638 +0000 UTC m=+158.372138165 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.397660 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:05 crc kubenswrapper[4818]: E1122 04:50:05.397840 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:05.897814229 +0000 UTC m=+158.472230756 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.398172 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:05 crc kubenswrapper[4818]: E1122 04:50:05.398497 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 04:50:05.898489977 +0000 UTC m=+158.472906504 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cd4qt" (UID: "677aa266-f94d-498a-8034-1205889c4958") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.441948 4818 patch_prober.go:28] interesting pod/router-default-5444994796-5gfmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 04:50:05 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Nov 22 04:50:05 crc kubenswrapper[4818]: [+]process-running ok Nov 22 04:50:05 crc kubenswrapper[4818]: healthz check failed Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.441999 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5gfmz" podUID="eb0cffc3-bf07-411c-822e-aa45c58387cb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.478102 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 22 04:50:05 crc kubenswrapper[4818]: W1122 04:50:05.489364 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod9b4aabc4_c336_4114_86f4_4f94bb336609.slice/crio-1e737a3ec7bbefc8298946fd1726ee0b1d82635dbc735b97fb467a4c7b362024 WatchSource:0}: Error finding container 1e737a3ec7bbefc8298946fd1726ee0b1d82635dbc735b97fb467a4c7b362024: Status 404 returned error can't find the container with id 1e737a3ec7bbefc8298946fd1726ee0b1d82635dbc735b97fb467a4c7b362024 Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.498885 4818 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-22T04:50:04.673026953Z","Handler":null,"Name":""} Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.500266 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:05 crc kubenswrapper[4818]: E1122 04:50:05.500589 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 04:50:06.000577561 +0000 UTC m=+158.574994088 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.509761 4818 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.509791 4818 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.601600 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.604367 4818 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.604408 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.642881 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cd4qt\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.675486 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9b4aabc4-c336-4114-86f4-4f94bb336609","Type":"ContainerStarted","Data":"1e737a3ec7bbefc8298946fd1726ee0b1d82635dbc735b97fb467a4c7b362024"} Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.677263 4818 generic.go:334] "Generic (PLEG): container finished" podID="1146dedf-aeff-4812-84e5-dcfe1cfc56ca" containerID="4104ec4e5e0aa69fa66743a91c89553a4bdfff3e7c692859b10f413333cadc3e" exitCode=0 Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.677303 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7spb" event={"ID":"1146dedf-aeff-4812-84e5-dcfe1cfc56ca","Type":"ContainerDied","Data":"4104ec4e5e0aa69fa66743a91c89553a4bdfff3e7c692859b10f413333cadc3e"} Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.677348 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7spb" event={"ID":"1146dedf-aeff-4812-84e5-dcfe1cfc56ca","Type":"ContainerStarted","Data":"cfcd480d80a31ea0921b3db61192d1a184c3c326afcb733df013a4fbd39fa753"} Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.678523 4818 generic.go:334] "Generic (PLEG): container finished" podID="30e236a6-d269-4f06-ae6e-e3c9bb74b4bf" containerID="4e09d0edf82e0b5cf5fdac97850852843dfac2b775e16e47c9025a816ecb4d62" exitCode=0 Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.678568 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg659" event={"ID":"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf","Type":"ContainerDied","Data":"4e09d0edf82e0b5cf5fdac97850852843dfac2b775e16e47c9025a816ecb4d62"} Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.678628 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg659" event={"ID":"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf","Type":"ContainerStarted","Data":"2648717bb6765491289df8b05a194e3c501ac1fcf258a4c38e8d3fc7f024f27a"} Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.679121 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.679917 4818 generic.go:334] "Generic (PLEG): container finished" podID="0e0614cc-57b5-4782-b462-e693c03c2117" containerID="bc26224b836a30445b253579c27f8308345be71519b52f86ecd93d4240ea27ff" exitCode=0 Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.679971 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blsms" event={"ID":"0e0614cc-57b5-4782-b462-e693c03c2117","Type":"ContainerDied","Data":"bc26224b836a30445b253579c27f8308345be71519b52f86ecd93d4240ea27ff"} Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.679986 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blsms" event={"ID":"0e0614cc-57b5-4782-b462-e693c03c2117","Type":"ContainerStarted","Data":"1568c4aee9967dab35f69933192e224daed6ade8e9de040b5fa5a275d1664f09"} Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.682144 4818 generic.go:334] "Generic (PLEG): container finished" podID="751d9838-4c34-4251-b3a6-1d808f98a034" containerID="b7ccf6acf2f43368232355bb645f785a164ce14e18e91b3babade5df387c1565" exitCode=0 Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.682244 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4" event={"ID":"751d9838-4c34-4251-b3a6-1d808f98a034","Type":"ContainerDied","Data":"b7ccf6acf2f43368232355bb645f785a164ce14e18e91b3babade5df387c1565"} Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.685927 4818 generic.go:334] "Generic (PLEG): container finished" podID="b5d474c0-a83d-4383-a4d9-6fd0e128d34b" containerID="8499a55582668439375b5171b8bdb3172c405b079e6e0d74af514f3f7d54b04c" exitCode=0 Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.686001 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xsv9c" event={"ID":"b5d474c0-a83d-4383-a4d9-6fd0e128d34b","Type":"ContainerDied","Data":"8499a55582668439375b5171b8bdb3172c405b079e6e0d74af514f3f7d54b04c"} Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.686027 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xsv9c" event={"ID":"b5d474c0-a83d-4383-a4d9-6fd0e128d34b","Type":"ContainerStarted","Data":"8f7cfcbd668cf10f2fa4d8b296d4584a206d1c4c5b32cb57e33f37e02cad47ee"} Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.689668 4818 generic.go:334] "Generic (PLEG): container finished" podID="5bf364b3-3f72-4ccb-8286-26f5eec4e399" containerID="e62b4358058a03d4a7e01f9eb796d0d9f43afa82f01f3d54862ff036983efbf8" exitCode=0 Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.689970 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5bf364b3-3f72-4ccb-8286-26f5eec4e399","Type":"ContainerDied","Data":"e62b4358058a03d4a7e01f9eb796d0d9f43afa82f01f3d54862ff036983efbf8"} Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.702648 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.757680 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.809049 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.920372 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jswlc"] Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.921301 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jswlc" Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.929362 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 22 04:50:05 crc kubenswrapper[4818]: I1122 04:50:05.943907 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jswlc"] Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.107303 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shjjf\" (UniqueName: \"kubernetes.io/projected/e9310d3a-0f3d-4b29-9ab6-ea9480f40274-kube-api-access-shjjf\") pod \"redhat-marketplace-jswlc\" (UID: \"e9310d3a-0f3d-4b29-9ab6-ea9480f40274\") " pod="openshift-marketplace/redhat-marketplace-jswlc" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.107698 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9310d3a-0f3d-4b29-9ab6-ea9480f40274-catalog-content\") pod \"redhat-marketplace-jswlc\" (UID: \"e9310d3a-0f3d-4b29-9ab6-ea9480f40274\") " pod="openshift-marketplace/redhat-marketplace-jswlc" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.107731 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9310d3a-0f3d-4b29-9ab6-ea9480f40274-utilities\") pod \"redhat-marketplace-jswlc\" (UID: \"e9310d3a-0f3d-4b29-9ab6-ea9480f40274\") " pod="openshift-marketplace/redhat-marketplace-jswlc" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.125533 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cd4qt"] Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.208427 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9310d3a-0f3d-4b29-9ab6-ea9480f40274-catalog-content\") pod \"redhat-marketplace-jswlc\" (UID: \"e9310d3a-0f3d-4b29-9ab6-ea9480f40274\") " pod="openshift-marketplace/redhat-marketplace-jswlc" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.208476 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9310d3a-0f3d-4b29-9ab6-ea9480f40274-utilities\") pod \"redhat-marketplace-jswlc\" (UID: \"e9310d3a-0f3d-4b29-9ab6-ea9480f40274\") " pod="openshift-marketplace/redhat-marketplace-jswlc" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.208560 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shjjf\" (UniqueName: \"kubernetes.io/projected/e9310d3a-0f3d-4b29-9ab6-ea9480f40274-kube-api-access-shjjf\") pod \"redhat-marketplace-jswlc\" (UID: \"e9310d3a-0f3d-4b29-9ab6-ea9480f40274\") " pod="openshift-marketplace/redhat-marketplace-jswlc" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.209049 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9310d3a-0f3d-4b29-9ab6-ea9480f40274-catalog-content\") pod \"redhat-marketplace-jswlc\" (UID: \"e9310d3a-0f3d-4b29-9ab6-ea9480f40274\") " pod="openshift-marketplace/redhat-marketplace-jswlc" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.209133 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9310d3a-0f3d-4b29-9ab6-ea9480f40274-utilities\") pod \"redhat-marketplace-jswlc\" (UID: \"e9310d3a-0f3d-4b29-9ab6-ea9480f40274\") " pod="openshift-marketplace/redhat-marketplace-jswlc" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.231051 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shjjf\" (UniqueName: \"kubernetes.io/projected/e9310d3a-0f3d-4b29-9ab6-ea9480f40274-kube-api-access-shjjf\") pod \"redhat-marketplace-jswlc\" (UID: \"e9310d3a-0f3d-4b29-9ab6-ea9480f40274\") " pod="openshift-marketplace/redhat-marketplace-jswlc" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.236566 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jswlc" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.297266 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.318284 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5n24q"] Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.319225 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5n24q" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.332122 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5n24q"] Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.429108 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jswlc"] Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.437200 4818 patch_prober.go:28] interesting pod/router-default-5444994796-5gfmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 04:50:06 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Nov 22 04:50:06 crc kubenswrapper[4818]: [+]process-running ok Nov 22 04:50:06 crc kubenswrapper[4818]: healthz check failed Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.437275 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5gfmz" podUID="eb0cffc3-bf07-411c-822e-aa45c58387cb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.511950 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p26q4\" (UniqueName: \"kubernetes.io/projected/618de3ee-2743-44e0-9d2d-b85b1397d056-kube-api-access-p26q4\") pod \"redhat-marketplace-5n24q\" (UID: \"618de3ee-2743-44e0-9d2d-b85b1397d056\") " pod="openshift-marketplace/redhat-marketplace-5n24q" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.512005 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/618de3ee-2743-44e0-9d2d-b85b1397d056-catalog-content\") pod \"redhat-marketplace-5n24q\" (UID: \"618de3ee-2743-44e0-9d2d-b85b1397d056\") " pod="openshift-marketplace/redhat-marketplace-5n24q" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.512072 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/618de3ee-2743-44e0-9d2d-b85b1397d056-utilities\") pod \"redhat-marketplace-5n24q\" (UID: \"618de3ee-2743-44e0-9d2d-b85b1397d056\") " pod="openshift-marketplace/redhat-marketplace-5n24q" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.534416 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.534455 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.541750 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.613601 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p26q4\" (UniqueName: \"kubernetes.io/projected/618de3ee-2743-44e0-9d2d-b85b1397d056-kube-api-access-p26q4\") pod \"redhat-marketplace-5n24q\" (UID: \"618de3ee-2743-44e0-9d2d-b85b1397d056\") " pod="openshift-marketplace/redhat-marketplace-5n24q" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.613659 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/618de3ee-2743-44e0-9d2d-b85b1397d056-catalog-content\") pod \"redhat-marketplace-5n24q\" (UID: \"618de3ee-2743-44e0-9d2d-b85b1397d056\") " pod="openshift-marketplace/redhat-marketplace-5n24q" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.613686 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/618de3ee-2743-44e0-9d2d-b85b1397d056-utilities\") pod \"redhat-marketplace-5n24q\" (UID: \"618de3ee-2743-44e0-9d2d-b85b1397d056\") " pod="openshift-marketplace/redhat-marketplace-5n24q" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.614143 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/618de3ee-2743-44e0-9d2d-b85b1397d056-catalog-content\") pod \"redhat-marketplace-5n24q\" (UID: \"618de3ee-2743-44e0-9d2d-b85b1397d056\") " pod="openshift-marketplace/redhat-marketplace-5n24q" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.614158 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/618de3ee-2743-44e0-9d2d-b85b1397d056-utilities\") pod \"redhat-marketplace-5n24q\" (UID: \"618de3ee-2743-44e0-9d2d-b85b1397d056\") " pod="openshift-marketplace/redhat-marketplace-5n24q" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.628398 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.636028 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p26q4\" (UniqueName: \"kubernetes.io/projected/618de3ee-2743-44e0-9d2d-b85b1397d056-kube-api-access-p26q4\") pod \"redhat-marketplace-5n24q\" (UID: \"618de3ee-2743-44e0-9d2d-b85b1397d056\") " pod="openshift-marketplace/redhat-marketplace-5n24q" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.642798 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.642853 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.642918 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.642976 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.648937 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5n24q" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.663334 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-44grv" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.706386 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jswlc" event={"ID":"e9310d3a-0f3d-4b29-9ab6-ea9480f40274","Type":"ContainerStarted","Data":"d2c01cd7b5ac31c71cd8423201f20a33b1f945ecba8213ad5506cf2b3290c40c"} Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.706443 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jswlc" event={"ID":"e9310d3a-0f3d-4b29-9ab6-ea9480f40274","Type":"ContainerStarted","Data":"bb765bd3f43771442f8485d583573f762bfa41288f71f3732719b7087f819eae"} Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.706464 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.706963 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.709775 4818 generic.go:334] "Generic (PLEG): container finished" podID="9b4aabc4-c336-4114-86f4-4f94bb336609" containerID="341a7516cae49e9c581a96578060dca6b59429451d135105191d5844ea8d3897" exitCode=0 Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.709837 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9b4aabc4-c336-4114-86f4-4f94bb336609","Type":"ContainerDied","Data":"341a7516cae49e9c581a96578060dca6b59429451d135105191d5844ea8d3897"} Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.714684 4818 patch_prober.go:28] interesting pod/console-f9d7485db-h56w2 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.714722 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-h56w2" podUID="670a3117-320e-477c-8c4b-a54fe5a4ae9c" containerName="console" probeResult="failure" output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.724391 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" event={"ID":"677aa266-f94d-498a-8034-1205889c4958","Type":"ContainerStarted","Data":"3d29507ef31e0257cdef51333c6adb2b79403040b5a6dda3dc6df109fe9fdb41"} Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.724430 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" event={"ID":"677aa266-f94d-498a-8034-1205889c4958","Type":"ContainerStarted","Data":"90651c0e4b27fe0f9c8b8aa42ec12e3e28d4c4d2485a6455f7d7d546e5bc133a"} Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.724589 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.732114 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4lh8z" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.781546 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" podStartSLOduration=135.781527109 podStartE2EDuration="2m15.781527109s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:50:06.778297392 +0000 UTC m=+159.352713939" watchObservedRunningTime="2025-11-22 04:50:06.781527109 +0000 UTC m=+159.355943636" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.897097 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.909923 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-9nsph" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.920749 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7xg9d"] Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.922300 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7xg9d" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.923910 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 22 04:50:06 crc kubenswrapper[4818]: I1122 04:50:06.983587 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7xg9d"] Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.030047 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a48b14cb-0897-4530-8f74-bee9352bcc44-utilities\") pod \"redhat-operators-7xg9d\" (UID: \"a48b14cb-0897-4530-8f74-bee9352bcc44\") " pod="openshift-marketplace/redhat-operators-7xg9d" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.030121 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcjvh\" (UniqueName: \"kubernetes.io/projected/a48b14cb-0897-4530-8f74-bee9352bcc44-kube-api-access-bcjvh\") pod \"redhat-operators-7xg9d\" (UID: \"a48b14cb-0897-4530-8f74-bee9352bcc44\") " pod="openshift-marketplace/redhat-operators-7xg9d" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.030141 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a48b14cb-0897-4530-8f74-bee9352bcc44-catalog-content\") pod \"redhat-operators-7xg9d\" (UID: \"a48b14cb-0897-4530-8f74-bee9352bcc44\") " pod="openshift-marketplace/redhat-operators-7xg9d" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.131277 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a48b14cb-0897-4530-8f74-bee9352bcc44-utilities\") pod \"redhat-operators-7xg9d\" (UID: \"a48b14cb-0897-4530-8f74-bee9352bcc44\") " pod="openshift-marketplace/redhat-operators-7xg9d" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.131363 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcjvh\" (UniqueName: \"kubernetes.io/projected/a48b14cb-0897-4530-8f74-bee9352bcc44-kube-api-access-bcjvh\") pod \"redhat-operators-7xg9d\" (UID: \"a48b14cb-0897-4530-8f74-bee9352bcc44\") " pod="openshift-marketplace/redhat-operators-7xg9d" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.131388 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a48b14cb-0897-4530-8f74-bee9352bcc44-catalog-content\") pod \"redhat-operators-7xg9d\" (UID: \"a48b14cb-0897-4530-8f74-bee9352bcc44\") " pod="openshift-marketplace/redhat-operators-7xg9d" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.131790 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a48b14cb-0897-4530-8f74-bee9352bcc44-utilities\") pod \"redhat-operators-7xg9d\" (UID: \"a48b14cb-0897-4530-8f74-bee9352bcc44\") " pod="openshift-marketplace/redhat-operators-7xg9d" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.132659 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a48b14cb-0897-4530-8f74-bee9352bcc44-catalog-content\") pod \"redhat-operators-7xg9d\" (UID: \"a48b14cb-0897-4530-8f74-bee9352bcc44\") " pod="openshift-marketplace/redhat-operators-7xg9d" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.140496 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.157644 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.170433 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcjvh\" (UniqueName: \"kubernetes.io/projected/a48b14cb-0897-4530-8f74-bee9352bcc44-kube-api-access-bcjvh\") pod \"redhat-operators-7xg9d\" (UID: \"a48b14cb-0897-4530-8f74-bee9352bcc44\") " pod="openshift-marketplace/redhat-operators-7xg9d" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.180468 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-btgsr" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.194397 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jlvsb" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.225387 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k27q8" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.259209 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.267713 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7xg9d" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.289464 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.314487 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9p4pn"] Nov 22 04:50:07 crc kubenswrapper[4818]: E1122 04:50:07.314727 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bf364b3-3f72-4ccb-8286-26f5eec4e399" containerName="pruner" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.314742 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bf364b3-3f72-4ccb-8286-26f5eec4e399" containerName="pruner" Nov 22 04:50:07 crc kubenswrapper[4818]: E1122 04:50:07.314758 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="751d9838-4c34-4251-b3a6-1d808f98a034" containerName="collect-profiles" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.314764 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="751d9838-4c34-4251-b3a6-1d808f98a034" containerName="collect-profiles" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.314863 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bf364b3-3f72-4ccb-8286-26f5eec4e399" containerName="pruner" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.314874 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="751d9838-4c34-4251-b3a6-1d808f98a034" containerName="collect-profiles" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.315628 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9p4pn" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.317281 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5n24q"] Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.324052 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9p4pn"] Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.335997 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5bf364b3-3f72-4ccb-8286-26f5eec4e399-kubelet-dir\") pod \"5bf364b3-3f72-4ccb-8286-26f5eec4e399\" (UID: \"5bf364b3-3f72-4ccb-8286-26f5eec4e399\") " Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.336066 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5bf364b3-3f72-4ccb-8286-26f5eec4e399-kube-api-access\") pod \"5bf364b3-3f72-4ccb-8286-26f5eec4e399\" (UID: \"5bf364b3-3f72-4ccb-8286-26f5eec4e399\") " Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.336102 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/751d9838-4c34-4251-b3a6-1d808f98a034-config-volume\") pod \"751d9838-4c34-4251-b3a6-1d808f98a034\" (UID: \"751d9838-4c34-4251-b3a6-1d808f98a034\") " Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.336193 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/751d9838-4c34-4251-b3a6-1d808f98a034-secret-volume\") pod \"751d9838-4c34-4251-b3a6-1d808f98a034\" (UID: \"751d9838-4c34-4251-b3a6-1d808f98a034\") " Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.336284 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzppq\" (UniqueName: \"kubernetes.io/projected/751d9838-4c34-4251-b3a6-1d808f98a034-kube-api-access-xzppq\") pod \"751d9838-4c34-4251-b3a6-1d808f98a034\" (UID: \"751d9838-4c34-4251-b3a6-1d808f98a034\") " Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.336682 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5bf364b3-3f72-4ccb-8286-26f5eec4e399-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5bf364b3-3f72-4ccb-8286-26f5eec4e399" (UID: "5bf364b3-3f72-4ccb-8286-26f5eec4e399"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.338274 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/751d9838-4c34-4251-b3a6-1d808f98a034-config-volume" (OuterVolumeSpecName: "config-volume") pod "751d9838-4c34-4251-b3a6-1d808f98a034" (UID: "751d9838-4c34-4251-b3a6-1d808f98a034"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.340604 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/751d9838-4c34-4251-b3a6-1d808f98a034-kube-api-access-xzppq" (OuterVolumeSpecName: "kube-api-access-xzppq") pod "751d9838-4c34-4251-b3a6-1d808f98a034" (UID: "751d9838-4c34-4251-b3a6-1d808f98a034"). InnerVolumeSpecName "kube-api-access-xzppq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.350676 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/751d9838-4c34-4251-b3a6-1d808f98a034-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "751d9838-4c34-4251-b3a6-1d808f98a034" (UID: "751d9838-4c34-4251-b3a6-1d808f98a034"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.351145 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bf364b3-3f72-4ccb-8286-26f5eec4e399-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5bf364b3-3f72-4ccb-8286-26f5eec4e399" (UID: "5bf364b3-3f72-4ccb-8286-26f5eec4e399"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.439581 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.439920 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7srn\" (UniqueName: \"kubernetes.io/projected/c6499ef8-b253-4710-8e30-9546db961552-kube-api-access-n7srn\") pod \"redhat-operators-9p4pn\" (UID: \"c6499ef8-b253-4710-8e30-9546db961552\") " pod="openshift-marketplace/redhat-operators-9p4pn" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.439961 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6499ef8-b253-4710-8e30-9546db961552-utilities\") pod \"redhat-operators-9p4pn\" (UID: \"c6499ef8-b253-4710-8e30-9546db961552\") " pod="openshift-marketplace/redhat-operators-9p4pn" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.441640 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6499ef8-b253-4710-8e30-9546db961552-catalog-content\") pod \"redhat-operators-9p4pn\" (UID: \"c6499ef8-b253-4710-8e30-9546db961552\") " pod="openshift-marketplace/redhat-operators-9p4pn" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.441682 4818 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/751d9838-4c34-4251-b3a6-1d808f98a034-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.441695 4818 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/751d9838-4c34-4251-b3a6-1d808f98a034-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.441707 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzppq\" (UniqueName: \"kubernetes.io/projected/751d9838-4c34-4251-b3a6-1d808f98a034-kube-api-access-xzppq\") on node \"crc\" DevicePath \"\"" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.441717 4818 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5bf364b3-3f72-4ccb-8286-26f5eec4e399-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.441727 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5bf364b3-3f72-4ccb-8286-26f5eec4e399-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.443553 4818 patch_prober.go:28] interesting pod/router-default-5444994796-5gfmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 04:50:07 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Nov 22 04:50:07 crc kubenswrapper[4818]: [+]process-running ok Nov 22 04:50:07 crc kubenswrapper[4818]: healthz check failed Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.443598 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5gfmz" podUID="eb0cffc3-bf07-411c-822e-aa45c58387cb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.560032 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6499ef8-b253-4710-8e30-9546db961552-catalog-content\") pod \"redhat-operators-9p4pn\" (UID: \"c6499ef8-b253-4710-8e30-9546db961552\") " pod="openshift-marketplace/redhat-operators-9p4pn" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.560114 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7srn\" (UniqueName: \"kubernetes.io/projected/c6499ef8-b253-4710-8e30-9546db961552-kube-api-access-n7srn\") pod \"redhat-operators-9p4pn\" (UID: \"c6499ef8-b253-4710-8e30-9546db961552\") " pod="openshift-marketplace/redhat-operators-9p4pn" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.560139 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6499ef8-b253-4710-8e30-9546db961552-utilities\") pod \"redhat-operators-9p4pn\" (UID: \"c6499ef8-b253-4710-8e30-9546db961552\") " pod="openshift-marketplace/redhat-operators-9p4pn" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.560617 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6499ef8-b253-4710-8e30-9546db961552-catalog-content\") pod \"redhat-operators-9p4pn\" (UID: \"c6499ef8-b253-4710-8e30-9546db961552\") " pod="openshift-marketplace/redhat-operators-9p4pn" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.560864 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6499ef8-b253-4710-8e30-9546db961552-utilities\") pod \"redhat-operators-9p4pn\" (UID: \"c6499ef8-b253-4710-8e30-9546db961552\") " pod="openshift-marketplace/redhat-operators-9p4pn" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.600847 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7srn\" (UniqueName: \"kubernetes.io/projected/c6499ef8-b253-4710-8e30-9546db961552-kube-api-access-n7srn\") pod \"redhat-operators-9p4pn\" (UID: \"c6499ef8-b253-4710-8e30-9546db961552\") " pod="openshift-marketplace/redhat-operators-9p4pn" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.634728 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9p4pn" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.753367 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7xg9d"] Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.756174 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4" event={"ID":"751d9838-4c34-4251-b3a6-1d808f98a034","Type":"ContainerDied","Data":"0161b648d121799219d7a2b33f36d4578ab0b0aee754d072863e7e4ebbbbbbea"} Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.756214 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0161b648d121799219d7a2b33f36d4578ab0b0aee754d072863e7e4ebbbbbbea" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.756307 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.760767 4818 generic.go:334] "Generic (PLEG): container finished" podID="618de3ee-2743-44e0-9d2d-b85b1397d056" containerID="e131704fbc499124d84a87e92561df99983f1b2b6e26d37ae1856073fca84532" exitCode=0 Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.760826 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5n24q" event={"ID":"618de3ee-2743-44e0-9d2d-b85b1397d056","Type":"ContainerDied","Data":"e131704fbc499124d84a87e92561df99983f1b2b6e26d37ae1856073fca84532"} Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.760852 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5n24q" event={"ID":"618de3ee-2743-44e0-9d2d-b85b1397d056","Type":"ContainerStarted","Data":"8b7542a2f23c54b8ef16cac0ed6dd7ff8cd9f382f83ac5f7032c2d3c026b9caf"} Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.774549 4818 generic.go:334] "Generic (PLEG): container finished" podID="e9310d3a-0f3d-4b29-9ab6-ea9480f40274" containerID="d2c01cd7b5ac31c71cd8423201f20a33b1f945ecba8213ad5506cf2b3290c40c" exitCode=0 Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.774623 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jswlc" event={"ID":"e9310d3a-0f3d-4b29-9ab6-ea9480f40274","Type":"ContainerDied","Data":"d2c01cd7b5ac31c71cd8423201f20a33b1f945ecba8213ad5506cf2b3290c40c"} Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.785742 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5bf364b3-3f72-4ccb-8286-26f5eec4e399","Type":"ContainerDied","Data":"eedef202131044ba5571ada38de4b57707aa2d68bfae4fa5ae4d696b2edfae62"} Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.785785 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eedef202131044ba5571ada38de4b57707aa2d68bfae4fa5ae4d696b2edfae62" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.786891 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 04:50:07 crc kubenswrapper[4818]: I1122 04:50:07.978568 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9p4pn"] Nov 22 04:50:07 crc kubenswrapper[4818]: W1122 04:50:07.980717 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6499ef8_b253_4710_8e30_9546db961552.slice/crio-2d806f2c5c3db0b2ff4894b07736e27b0aaf0fd425d80773189028ed97a31874 WatchSource:0}: Error finding container 2d806f2c5c3db0b2ff4894b07736e27b0aaf0fd425d80773189028ed97a31874: Status 404 returned error can't find the container with id 2d806f2c5c3db0b2ff4894b07736e27b0aaf0fd425d80773189028ed97a31874 Nov 22 04:50:08 crc kubenswrapper[4818]: I1122 04:50:08.112408 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 04:50:08 crc kubenswrapper[4818]: I1122 04:50:08.173424 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9b4aabc4-c336-4114-86f4-4f94bb336609-kubelet-dir\") pod \"9b4aabc4-c336-4114-86f4-4f94bb336609\" (UID: \"9b4aabc4-c336-4114-86f4-4f94bb336609\") " Nov 22 04:50:08 crc kubenswrapper[4818]: I1122 04:50:08.173512 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b4aabc4-c336-4114-86f4-4f94bb336609-kube-api-access\") pod \"9b4aabc4-c336-4114-86f4-4f94bb336609\" (UID: \"9b4aabc4-c336-4114-86f4-4f94bb336609\") " Nov 22 04:50:08 crc kubenswrapper[4818]: I1122 04:50:08.174389 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b4aabc4-c336-4114-86f4-4f94bb336609-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9b4aabc4-c336-4114-86f4-4f94bb336609" (UID: "9b4aabc4-c336-4114-86f4-4f94bb336609"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 04:50:08 crc kubenswrapper[4818]: I1122 04:50:08.179534 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b4aabc4-c336-4114-86f4-4f94bb336609-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9b4aabc4-c336-4114-86f4-4f94bb336609" (UID: "9b4aabc4-c336-4114-86f4-4f94bb336609"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:50:08 crc kubenswrapper[4818]: I1122 04:50:08.275329 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b4aabc4-c336-4114-86f4-4f94bb336609-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 04:50:08 crc kubenswrapper[4818]: I1122 04:50:08.275698 4818 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9b4aabc4-c336-4114-86f4-4f94bb336609-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 22 04:50:08 crc kubenswrapper[4818]: I1122 04:50:08.298551 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 04:50:08 crc kubenswrapper[4818]: I1122 04:50:08.438456 4818 patch_prober.go:28] interesting pod/router-default-5444994796-5gfmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 04:50:08 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Nov 22 04:50:08 crc kubenswrapper[4818]: [+]process-running ok Nov 22 04:50:08 crc kubenswrapper[4818]: healthz check failed Nov 22 04:50:08 crc kubenswrapper[4818]: I1122 04:50:08.438515 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5gfmz" podUID="eb0cffc3-bf07-411c-822e-aa45c58387cb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 04:50:08 crc kubenswrapper[4818]: I1122 04:50:08.795289 4818 generic.go:334] "Generic (PLEG): container finished" podID="a48b14cb-0897-4530-8f74-bee9352bcc44" containerID="6750ee30fbe98f95438522cf05d598dff7b0e291eb14575aa75748e4e558cbfe" exitCode=0 Nov 22 04:50:08 crc kubenswrapper[4818]: I1122 04:50:08.795611 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xg9d" event={"ID":"a48b14cb-0897-4530-8f74-bee9352bcc44","Type":"ContainerDied","Data":"6750ee30fbe98f95438522cf05d598dff7b0e291eb14575aa75748e4e558cbfe"} Nov 22 04:50:08 crc kubenswrapper[4818]: I1122 04:50:08.795675 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xg9d" event={"ID":"a48b14cb-0897-4530-8f74-bee9352bcc44","Type":"ContainerStarted","Data":"60e27568ebe4434e6e4270b1d6154ccac213e437223ca96fd461627d16d5b119"} Nov 22 04:50:08 crc kubenswrapper[4818]: I1122 04:50:08.798156 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 04:50:08 crc kubenswrapper[4818]: I1122 04:50:08.798150 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9b4aabc4-c336-4114-86f4-4f94bb336609","Type":"ContainerDied","Data":"1e737a3ec7bbefc8298946fd1726ee0b1d82635dbc735b97fb467a4c7b362024"} Nov 22 04:50:08 crc kubenswrapper[4818]: I1122 04:50:08.798261 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e737a3ec7bbefc8298946fd1726ee0b1d82635dbc735b97fb467a4c7b362024" Nov 22 04:50:08 crc kubenswrapper[4818]: I1122 04:50:08.801717 4818 generic.go:334] "Generic (PLEG): container finished" podID="c6499ef8-b253-4710-8e30-9546db961552" containerID="7bd756d1b02621137e66144ce97af945367c246176629efd9f82e207a32441ee" exitCode=0 Nov 22 04:50:08 crc kubenswrapper[4818]: I1122 04:50:08.801773 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9p4pn" event={"ID":"c6499ef8-b253-4710-8e30-9546db961552","Type":"ContainerDied","Data":"7bd756d1b02621137e66144ce97af945367c246176629efd9f82e207a32441ee"} Nov 22 04:50:08 crc kubenswrapper[4818]: I1122 04:50:08.801809 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9p4pn" event={"ID":"c6499ef8-b253-4710-8e30-9546db961552","Type":"ContainerStarted","Data":"2d806f2c5c3db0b2ff4894b07736e27b0aaf0fd425d80773189028ed97a31874"} Nov 22 04:50:09 crc kubenswrapper[4818]: I1122 04:50:09.437267 4818 patch_prober.go:28] interesting pod/router-default-5444994796-5gfmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 04:50:09 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Nov 22 04:50:09 crc kubenswrapper[4818]: [+]process-running ok Nov 22 04:50:09 crc kubenswrapper[4818]: healthz check failed Nov 22 04:50:09 crc kubenswrapper[4818]: I1122 04:50:09.437321 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5gfmz" podUID="eb0cffc3-bf07-411c-822e-aa45c58387cb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 04:50:10 crc kubenswrapper[4818]: I1122 04:50:10.438637 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:50:10 crc kubenswrapper[4818]: I1122 04:50:10.441996 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-5gfmz" Nov 22 04:50:12 crc kubenswrapper[4818]: I1122 04:50:12.851719 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-b86wj" Nov 22 04:50:14 crc kubenswrapper[4818]: I1122 04:50:14.460492 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs\") pod \"network-metrics-daemon-djd77\" (UID: \"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\") " pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:50:14 crc kubenswrapper[4818]: I1122 04:50:14.468294 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/429a37e7-cb52-4e76-89ee-16f2b9e6fbfe-metrics-certs\") pod \"network-metrics-daemon-djd77\" (UID: \"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe\") " pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:50:14 crc kubenswrapper[4818]: I1122 04:50:14.544338 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-djd77" Nov 22 04:50:16 crc kubenswrapper[4818]: I1122 04:50:16.641228 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:50:16 crc kubenswrapper[4818]: I1122 04:50:16.641646 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:50:16 crc kubenswrapper[4818]: I1122 04:50:16.641406 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:50:16 crc kubenswrapper[4818]: I1122 04:50:16.641835 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:50:16 crc kubenswrapper[4818]: I1122 04:50:16.706559 4818 patch_prober.go:28] interesting pod/console-f9d7485db-h56w2 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Nov 22 04:50:16 crc kubenswrapper[4818]: I1122 04:50:16.706625 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-h56w2" podUID="670a3117-320e-477c-8c4b-a54fe5a4ae9c" containerName="console" probeResult="failure" output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" Nov 22 04:50:21 crc kubenswrapper[4818]: I1122 04:50:21.265243 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 04:50:21 crc kubenswrapper[4818]: I1122 04:50:21.265740 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 04:50:26 crc kubenswrapper[4818]: I1122 04:50:26.479928 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:50:26 crc kubenswrapper[4818]: I1122 04:50:26.642458 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:50:26 crc kubenswrapper[4818]: I1122 04:50:26.642530 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:50:26 crc kubenswrapper[4818]: I1122 04:50:26.642634 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:50:26 crc kubenswrapper[4818]: I1122 04:50:26.642695 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:50:26 crc kubenswrapper[4818]: I1122 04:50:26.642743 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-rcvlq" Nov 22 04:50:26 crc kubenswrapper[4818]: I1122 04:50:26.643408 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"4e47b7e752719f7b0e5755e3fa9f73a4fb266ff462c4ce5a897145e2e88c3f56"} pod="openshift-console/downloads-7954f5f757-rcvlq" containerMessage="Container download-server failed liveness probe, will be restarted" Nov 22 04:50:26 crc kubenswrapper[4818]: I1122 04:50:26.643501 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" containerID="cri-o://4e47b7e752719f7b0e5755e3fa9f73a4fb266ff462c4ce5a897145e2e88c3f56" gracePeriod=2 Nov 22 04:50:26 crc kubenswrapper[4818]: I1122 04:50:26.643576 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:50:26 crc kubenswrapper[4818]: I1122 04:50:26.643652 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:50:27 crc kubenswrapper[4818]: I1122 04:50:27.308382 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:50:27 crc kubenswrapper[4818]: I1122 04:50:27.315751 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-h56w2" Nov 22 04:50:28 crc kubenswrapper[4818]: I1122 04:50:28.476487 4818 patch_prober.go:28] interesting pod/router-default-5444994796-5gfmz container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 04:50:28 crc kubenswrapper[4818]: I1122 04:50:28.476939 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-5gfmz" podUID="eb0cffc3-bf07-411c-822e-aa45c58387cb" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 04:50:33 crc kubenswrapper[4818]: I1122 04:50:33.022605 4818 generic.go:334] "Generic (PLEG): container finished" podID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerID="4e47b7e752719f7b0e5755e3fa9f73a4fb266ff462c4ce5a897145e2e88c3f56" exitCode=0 Nov 22 04:50:33 crc kubenswrapper[4818]: I1122 04:50:33.022684 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-rcvlq" event={"ID":"61feab0c-5f50-49fe-bf35-de2a5f66bf32","Type":"ContainerDied","Data":"4e47b7e752719f7b0e5755e3fa9f73a4fb266ff462c4ce5a897145e2e88c3f56"} Nov 22 04:50:36 crc kubenswrapper[4818]: I1122 04:50:36.641710 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:50:36 crc kubenswrapper[4818]: I1122 04:50:36.642406 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:50:37 crc kubenswrapper[4818]: I1122 04:50:37.512850 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqnns" Nov 22 04:50:41 crc kubenswrapper[4818]: I1122 04:50:41.806939 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 04:50:46 crc kubenswrapper[4818]: I1122 04:50:46.642677 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:50:46 crc kubenswrapper[4818]: I1122 04:50:46.643133 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:50:51 crc kubenswrapper[4818]: I1122 04:50:51.265659 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 04:50:51 crc kubenswrapper[4818]: I1122 04:50:51.268732 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 04:50:51 crc kubenswrapper[4818]: E1122 04:50:51.975998 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 22 04:50:51 crc kubenswrapper[4818]: E1122 04:50:51.976761 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bclj5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-blsms_openshift-marketplace(0e0614cc-57b5-4782-b462-e693c03c2117): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 04:50:51 crc kubenswrapper[4818]: E1122 04:50:51.978170 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-blsms" podUID="0e0614cc-57b5-4782-b462-e693c03c2117" Nov 22 04:50:56 crc kubenswrapper[4818]: I1122 04:50:56.641174 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:50:56 crc kubenswrapper[4818]: I1122 04:50:56.641630 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:50:59 crc kubenswrapper[4818]: E1122 04:50:59.509316 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-blsms" podUID="0e0614cc-57b5-4782-b462-e693c03c2117" Nov 22 04:51:02 crc kubenswrapper[4818]: E1122 04:51:02.493699 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 22 04:51:02 crc kubenswrapper[4818]: E1122 04:51:02.493946 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kkf9w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-p7spb_openshift-marketplace(1146dedf-aeff-4812-84e5-dcfe1cfc56ca): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 04:51:02 crc kubenswrapper[4818]: E1122 04:51:02.495647 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-p7spb" podUID="1146dedf-aeff-4812-84e5-dcfe1cfc56ca" Nov 22 04:51:04 crc kubenswrapper[4818]: E1122 04:51:04.617209 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 22 04:51:04 crc kubenswrapper[4818]: E1122 04:51:04.617761 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-stfxm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-pg659_openshift-marketplace(30e236a6-d269-4f06-ae6e-e3c9bb74b4bf): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 04:51:04 crc kubenswrapper[4818]: E1122 04:51:04.618981 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-pg659" podUID="30e236a6-d269-4f06-ae6e-e3c9bb74b4bf" Nov 22 04:51:06 crc kubenswrapper[4818]: I1122 04:51:06.641364 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:51:06 crc kubenswrapper[4818]: I1122 04:51:06.641853 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:51:16 crc kubenswrapper[4818]: I1122 04:51:16.641082 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:51:16 crc kubenswrapper[4818]: I1122 04:51:16.641548 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:51:21 crc kubenswrapper[4818]: I1122 04:51:21.265521 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 04:51:21 crc kubenswrapper[4818]: I1122 04:51:21.266348 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 04:51:21 crc kubenswrapper[4818]: I1122 04:51:21.266441 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 04:51:21 crc kubenswrapper[4818]: I1122 04:51:21.267592 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863"} pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 04:51:21 crc kubenswrapper[4818]: I1122 04:51:21.267723 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" containerID="cri-o://2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863" gracePeriod=600 Nov 22 04:51:26 crc kubenswrapper[4818]: I1122 04:51:26.641717 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:51:26 crc kubenswrapper[4818]: I1122 04:51:26.642430 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:51:30 crc kubenswrapper[4818]: E1122 04:51:30.329491 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 22 04:51:30 crc kubenswrapper[4818]: E1122 04:51:30.329671 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lggqt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-xsv9c_openshift-marketplace(b5d474c0-a83d-4383-a4d9-6fd0e128d34b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 04:51:30 crc kubenswrapper[4818]: E1122 04:51:30.330889 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-xsv9c" podUID="b5d474c0-a83d-4383-a4d9-6fd0e128d34b" Nov 22 04:51:31 crc kubenswrapper[4818]: E1122 04:51:31.826578 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsv9c" podUID="b5d474c0-a83d-4383-a4d9-6fd0e128d34b" Nov 22 04:51:36 crc kubenswrapper[4818]: I1122 04:51:36.641246 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:51:36 crc kubenswrapper[4818]: I1122 04:51:36.644003 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:51:41 crc kubenswrapper[4818]: E1122 04:51:41.961416 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 22 04:51:41 crc kubenswrapper[4818]: E1122 04:51:41.961805 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-shjjf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-jswlc_openshift-marketplace(e9310d3a-0f3d-4b29-9ab6-ea9480f40274): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 04:51:41 crc kubenswrapper[4818]: E1122 04:51:41.962966 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-jswlc" podUID="e9310d3a-0f3d-4b29-9ab6-ea9480f40274" Nov 22 04:51:46 crc kubenswrapper[4818]: I1122 04:51:46.641319 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:51:46 crc kubenswrapper[4818]: I1122 04:51:46.641774 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:51:49 crc kubenswrapper[4818]: E1122 04:51:49.336983 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-jswlc" podUID="e9310d3a-0f3d-4b29-9ab6-ea9480f40274" Nov 22 04:51:49 crc kubenswrapper[4818]: I1122 04:51:49.599826 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-djd77"] Nov 22 04:51:50 crc kubenswrapper[4818]: E1122 04:51:50.025585 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 22 04:51:50 crc kubenswrapper[4818]: E1122 04:51:50.025764 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n7srn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-9p4pn_openshift-marketplace(c6499ef8-b253-4710-8e30-9546db961552): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 04:51:50 crc kubenswrapper[4818]: E1122 04:51:50.026945 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-9p4pn" podUID="c6499ef8-b253-4710-8e30-9546db961552" Nov 22 04:51:50 crc kubenswrapper[4818]: I1122 04:51:50.540602 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-djd77" event={"ID":"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe","Type":"ContainerStarted","Data":"6127acf0f07f889c3d20885c2f6c8b2c3e19a392098e8ae6c5862b2289ca41d1"} Nov 22 04:51:50 crc kubenswrapper[4818]: I1122 04:51:50.543717 4818 generic.go:334] "Generic (PLEG): container finished" podID="0971c9ec-29ce-44eb-9875-ad4664143305" containerID="2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863" exitCode=0 Nov 22 04:51:50 crc kubenswrapper[4818]: I1122 04:51:50.543840 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerDied","Data":"2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863"} Nov 22 04:51:54 crc kubenswrapper[4818]: E1122 04:51:54.365694 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 22 04:51:54 crc kubenswrapper[4818]: E1122 04:51:54.366183 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p26q4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-5n24q_openshift-marketplace(618de3ee-2743-44e0-9d2d-b85b1397d056): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 04:51:54 crc kubenswrapper[4818]: E1122 04:51:54.367529 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-5n24q" podUID="618de3ee-2743-44e0-9d2d-b85b1397d056" Nov 22 04:51:54 crc kubenswrapper[4818]: E1122 04:51:54.548500 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 22 04:51:54 crc kubenswrapper[4818]: E1122 04:51:54.548735 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bcjvh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-7xg9d_openshift-marketplace(a48b14cb-0897-4530-8f74-bee9352bcc44): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 04:51:54 crc kubenswrapper[4818]: E1122 04:51:54.549984 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-7xg9d" podUID="a48b14cb-0897-4530-8f74-bee9352bcc44" Nov 22 04:51:56 crc kubenswrapper[4818]: I1122 04:51:56.641625 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:51:56 crc kubenswrapper[4818]: I1122 04:51:56.642047 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:52:06 crc kubenswrapper[4818]: I1122 04:52:06.641184 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:52:06 crc kubenswrapper[4818]: I1122 04:52:06.642626 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:52:10 crc kubenswrapper[4818]: I1122 04:52:10.668930 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-rcvlq" event={"ID":"61feab0c-5f50-49fe-bf35-de2a5f66bf32","Type":"ContainerStarted","Data":"909a465e1dd06e52fcf4caa34a0e540fd0b4822da18c46a2be9b7c47e6174f06"} Nov 22 04:52:11 crc kubenswrapper[4818]: I1122 04:52:11.675049 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-djd77" event={"ID":"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe","Type":"ContainerStarted","Data":"e471762e8e311943186b18ad3a9e563ac4e32ff12f044e15437b1fa606e1fc3f"} Nov 22 04:52:11 crc kubenswrapper[4818]: I1122 04:52:11.676996 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xsv9c" event={"ID":"b5d474c0-a83d-4383-a4d9-6fd0e128d34b","Type":"ContainerStarted","Data":"8a90bc14671a07eadef9a712e2c4e9b9276f9543f2512b4f68c64b6bb097c014"} Nov 22 04:52:11 crc kubenswrapper[4818]: I1122 04:52:11.678877 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"57a015280b24f5d226048157fcf5b764a30fd55452acd0f6fc61ec5f33defeca"} Nov 22 04:52:11 crc kubenswrapper[4818]: I1122 04:52:11.688288 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7spb" event={"ID":"1146dedf-aeff-4812-84e5-dcfe1cfc56ca","Type":"ContainerStarted","Data":"3a15bf524b1d9f4914e4ad0b1a97fce97e8e952dc7c7bb977cd258cff18dd7c3"} Nov 22 04:52:11 crc kubenswrapper[4818]: I1122 04:52:11.693320 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg659" event={"ID":"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf","Type":"ContainerStarted","Data":"d7d124f9a3f300c251e85323f4282b72f96feeda714da160d9fd4a426c3f5e0e"} Nov 22 04:52:11 crc kubenswrapper[4818]: I1122 04:52:11.695120 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blsms" event={"ID":"0e0614cc-57b5-4782-b462-e693c03c2117","Type":"ContainerStarted","Data":"e8da1f9fbc283ea63d1501ce1f463e25a8f883ba1be36bf9cc7cf0f7f095aee2"} Nov 22 04:52:11 crc kubenswrapper[4818]: I1122 04:52:11.696171 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-rcvlq" Nov 22 04:52:11 crc kubenswrapper[4818]: I1122 04:52:11.696390 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:52:11 crc kubenswrapper[4818]: I1122 04:52:11.696435 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:52:12 crc kubenswrapper[4818]: I1122 04:52:12.705444 4818 generic.go:334] "Generic (PLEG): container finished" podID="1146dedf-aeff-4812-84e5-dcfe1cfc56ca" containerID="3a15bf524b1d9f4914e4ad0b1a97fce97e8e952dc7c7bb977cd258cff18dd7c3" exitCode=0 Nov 22 04:52:12 crc kubenswrapper[4818]: I1122 04:52:12.705586 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7spb" event={"ID":"1146dedf-aeff-4812-84e5-dcfe1cfc56ca","Type":"ContainerDied","Data":"3a15bf524b1d9f4914e4ad0b1a97fce97e8e952dc7c7bb977cd258cff18dd7c3"} Nov 22 04:52:12 crc kubenswrapper[4818]: I1122 04:52:12.708368 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg659" event={"ID":"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf","Type":"ContainerDied","Data":"d7d124f9a3f300c251e85323f4282b72f96feeda714da160d9fd4a426c3f5e0e"} Nov 22 04:52:12 crc kubenswrapper[4818]: I1122 04:52:12.708197 4818 generic.go:334] "Generic (PLEG): container finished" podID="30e236a6-d269-4f06-ae6e-e3c9bb74b4bf" containerID="d7d124f9a3f300c251e85323f4282b72f96feeda714da160d9fd4a426c3f5e0e" exitCode=0 Nov 22 04:52:12 crc kubenswrapper[4818]: I1122 04:52:12.712122 4818 generic.go:334] "Generic (PLEG): container finished" podID="0e0614cc-57b5-4782-b462-e693c03c2117" containerID="e8da1f9fbc283ea63d1501ce1f463e25a8f883ba1be36bf9cc7cf0f7f095aee2" exitCode=0 Nov 22 04:52:12 crc kubenswrapper[4818]: I1122 04:52:12.712239 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blsms" event={"ID":"0e0614cc-57b5-4782-b462-e693c03c2117","Type":"ContainerDied","Data":"e8da1f9fbc283ea63d1501ce1f463e25a8f883ba1be36bf9cc7cf0f7f095aee2"} Nov 22 04:52:12 crc kubenswrapper[4818]: I1122 04:52:12.715013 4818 generic.go:334] "Generic (PLEG): container finished" podID="b5d474c0-a83d-4383-a4d9-6fd0e128d34b" containerID="8a90bc14671a07eadef9a712e2c4e9b9276f9543f2512b4f68c64b6bb097c014" exitCode=0 Nov 22 04:52:12 crc kubenswrapper[4818]: I1122 04:52:12.715148 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xsv9c" event={"ID":"b5d474c0-a83d-4383-a4d9-6fd0e128d34b","Type":"ContainerDied","Data":"8a90bc14671a07eadef9a712e2c4e9b9276f9543f2512b4f68c64b6bb097c014"} Nov 22 04:52:12 crc kubenswrapper[4818]: I1122 04:52:12.715750 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:52:12 crc kubenswrapper[4818]: I1122 04:52:12.715799 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:52:13 crc kubenswrapper[4818]: I1122 04:52:13.723635 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-djd77" event={"ID":"429a37e7-cb52-4e76-89ee-16f2b9e6fbfe","Type":"ContainerStarted","Data":"30f3c10d76d634f47265ccf0fdc82a4153fd86d647c772d39c701bc47e66fc48"} Nov 22 04:52:13 crc kubenswrapper[4818]: I1122 04:52:13.744754 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-djd77" podStartSLOduration=262.744734169 podStartE2EDuration="4m22.744734169s" podCreationTimestamp="2025-11-22 04:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:52:13.74444096 +0000 UTC m=+286.318857487" watchObservedRunningTime="2025-11-22 04:52:13.744734169 +0000 UTC m=+286.319150696" Nov 22 04:52:16 crc kubenswrapper[4818]: I1122 04:52:16.640905 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:52:16 crc kubenswrapper[4818]: I1122 04:52:16.641407 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:52:16 crc kubenswrapper[4818]: I1122 04:52:16.640988 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:52:16 crc kubenswrapper[4818]: I1122 04:52:16.641555 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:52:26 crc kubenswrapper[4818]: I1122 04:52:26.641879 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:52:26 crc kubenswrapper[4818]: I1122 04:52:26.641912 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcvlq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 04:52:26 crc kubenswrapper[4818]: I1122 04:52:26.642762 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:52:26 crc kubenswrapper[4818]: I1122 04:52:26.642850 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcvlq" podUID="61feab0c-5f50-49fe-bf35-de2a5f66bf32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 04:52:28 crc kubenswrapper[4818]: I1122 04:52:28.025534 4818 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Nov 22 04:52:36 crc kubenswrapper[4818]: I1122 04:52:36.654326 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-rcvlq" Nov 22 04:53:33 crc kubenswrapper[4818]: I1122 04:53:33.367810 4818 generic.go:334] "Generic (PLEG): container finished" podID="c6499ef8-b253-4710-8e30-9546db961552" containerID="0d40aadf30061933224d26c5bc9dfd8641cd22b347bd83e655cfde85f5d25412" exitCode=0 Nov 22 04:53:33 crc kubenswrapper[4818]: I1122 04:53:33.367918 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9p4pn" event={"ID":"c6499ef8-b253-4710-8e30-9546db961552","Type":"ContainerDied","Data":"0d40aadf30061933224d26c5bc9dfd8641cd22b347bd83e655cfde85f5d25412"} Nov 22 04:53:33 crc kubenswrapper[4818]: I1122 04:53:33.372094 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7spb" event={"ID":"1146dedf-aeff-4812-84e5-dcfe1cfc56ca","Type":"ContainerStarted","Data":"66f2f515a6668df57bfe3138c5246f9ef9e6362092b0143e711b709080abe811"} Nov 22 04:53:33 crc kubenswrapper[4818]: I1122 04:53:33.375511 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg659" event={"ID":"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf","Type":"ContainerStarted","Data":"b00ce629adadad84a158329e4f6d5007a908430dbc8e874a6cc60a17b73dbdc2"} Nov 22 04:53:33 crc kubenswrapper[4818]: I1122 04:53:33.379620 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blsms" event={"ID":"0e0614cc-57b5-4782-b462-e693c03c2117","Type":"ContainerStarted","Data":"d57a8171c3e97f6c8a2297c4ba361e5d01f7fa77f4fbcb6d72a3d2428031fa78"} Nov 22 04:53:33 crc kubenswrapper[4818]: I1122 04:53:33.382471 4818 generic.go:334] "Generic (PLEG): container finished" podID="a48b14cb-0897-4530-8f74-bee9352bcc44" containerID="4b0cbc23059764fad28d1e357825d9bb4c667b16a2fdae1c3c2828111669d1c5" exitCode=0 Nov 22 04:53:33 crc kubenswrapper[4818]: I1122 04:53:33.382520 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xg9d" event={"ID":"a48b14cb-0897-4530-8f74-bee9352bcc44","Type":"ContainerDied","Data":"4b0cbc23059764fad28d1e357825d9bb4c667b16a2fdae1c3c2828111669d1c5"} Nov 22 04:53:33 crc kubenswrapper[4818]: I1122 04:53:33.385652 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xsv9c" event={"ID":"b5d474c0-a83d-4383-a4d9-6fd0e128d34b","Type":"ContainerStarted","Data":"c2acf0ae27cecbf073af179865e66acb6ebcf25552d30adce84399fb8fcd2ffa"} Nov 22 04:53:33 crc kubenswrapper[4818]: I1122 04:53:33.389091 4818 generic.go:334] "Generic (PLEG): container finished" podID="618de3ee-2743-44e0-9d2d-b85b1397d056" containerID="2ee78d0334a79dce9184e6181b232555a3e2863948649a205067236c72d4ed84" exitCode=0 Nov 22 04:53:33 crc kubenswrapper[4818]: I1122 04:53:33.389474 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5n24q" event={"ID":"618de3ee-2743-44e0-9d2d-b85b1397d056","Type":"ContainerDied","Data":"2ee78d0334a79dce9184e6181b232555a3e2863948649a205067236c72d4ed84"} Nov 22 04:53:33 crc kubenswrapper[4818]: I1122 04:53:33.391778 4818 generic.go:334] "Generic (PLEG): container finished" podID="e9310d3a-0f3d-4b29-9ab6-ea9480f40274" containerID="523bb3e9efd24499c8c60fdbbdd08f33ed2d49d3e16128898f98afd11cb1777f" exitCode=0 Nov 22 04:53:33 crc kubenswrapper[4818]: I1122 04:53:33.391837 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jswlc" event={"ID":"e9310d3a-0f3d-4b29-9ab6-ea9480f40274","Type":"ContainerDied","Data":"523bb3e9efd24499c8c60fdbbdd08f33ed2d49d3e16128898f98afd11cb1777f"} Nov 22 04:53:33 crc kubenswrapper[4818]: I1122 04:53:33.424444 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-blsms" podStartSLOduration=8.630315188 podStartE2EDuration="3m30.42441701s" podCreationTimestamp="2025-11-22 04:50:03 +0000 UTC" firstStartedPulling="2025-11-22 04:50:05.68157729 +0000 UTC m=+158.255993817" lastFinishedPulling="2025-11-22 04:53:27.475679072 +0000 UTC m=+360.050095639" observedRunningTime="2025-11-22 04:53:33.420792173 +0000 UTC m=+365.995208710" watchObservedRunningTime="2025-11-22 04:53:33.42441701 +0000 UTC m=+365.998833547" Nov 22 04:53:33 crc kubenswrapper[4818]: I1122 04:53:33.470586 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p7spb" podStartSLOduration=4.3594834129999995 podStartE2EDuration="3m29.470561279s" podCreationTimestamp="2025-11-22 04:50:04 +0000 UTC" firstStartedPulling="2025-11-22 04:50:05.678876557 +0000 UTC m=+158.253293084" lastFinishedPulling="2025-11-22 04:53:30.789954393 +0000 UTC m=+363.364370950" observedRunningTime="2025-11-22 04:53:33.466731726 +0000 UTC m=+366.041148283" watchObservedRunningTime="2025-11-22 04:53:33.470561279 +0000 UTC m=+366.044977836" Nov 22 04:53:33 crc kubenswrapper[4818]: I1122 04:53:33.486599 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xsv9c" podStartSLOduration=5.6140480010000005 podStartE2EDuration="3m29.486560539s" podCreationTimestamp="2025-11-22 04:50:04 +0000 UTC" firstStartedPulling="2025-11-22 04:50:05.687396028 +0000 UTC m=+158.261812555" lastFinishedPulling="2025-11-22 04:53:29.559908536 +0000 UTC m=+362.134325093" observedRunningTime="2025-11-22 04:53:33.484803822 +0000 UTC m=+366.059220389" watchObservedRunningTime="2025-11-22 04:53:33.486560539 +0000 UTC m=+366.060977066" Nov 22 04:53:33 crc kubenswrapper[4818]: I1122 04:53:33.540247 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pg659" podStartSLOduration=8.743756063 podStartE2EDuration="3m30.54022135s" podCreationTimestamp="2025-11-22 04:50:03 +0000 UTC" firstStartedPulling="2025-11-22 04:50:05.679937085 +0000 UTC m=+158.254353602" lastFinishedPulling="2025-11-22 04:53:27.476402312 +0000 UTC m=+360.050818889" observedRunningTime="2025-11-22 04:53:33.539047288 +0000 UTC m=+366.113463865" watchObservedRunningTime="2025-11-22 04:53:33.54022135 +0000 UTC m=+366.114637907" Nov 22 04:53:34 crc kubenswrapper[4818]: I1122 04:53:34.051635 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pg659" Nov 22 04:53:34 crc kubenswrapper[4818]: I1122 04:53:34.051981 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pg659" Nov 22 04:53:34 crc kubenswrapper[4818]: I1122 04:53:34.238144 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-blsms" Nov 22 04:53:34 crc kubenswrapper[4818]: I1122 04:53:34.238333 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-blsms" Nov 22 04:53:34 crc kubenswrapper[4818]: I1122 04:53:34.482825 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xsv9c" Nov 22 04:53:34 crc kubenswrapper[4818]: I1122 04:53:34.482890 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xsv9c" Nov 22 04:53:34 crc kubenswrapper[4818]: I1122 04:53:34.685582 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p7spb" Nov 22 04:53:34 crc kubenswrapper[4818]: I1122 04:53:34.685624 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p7spb" Nov 22 04:53:35 crc kubenswrapper[4818]: I1122 04:53:35.817872 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-blsms" podUID="0e0614cc-57b5-4782-b462-e693c03c2117" containerName="registry-server" probeResult="failure" output=< Nov 22 04:53:35 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Nov 22 04:53:35 crc kubenswrapper[4818]: > Nov 22 04:53:35 crc kubenswrapper[4818]: I1122 04:53:35.820189 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-xsv9c" podUID="b5d474c0-a83d-4383-a4d9-6fd0e128d34b" containerName="registry-server" probeResult="failure" output=< Nov 22 04:53:35 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Nov 22 04:53:35 crc kubenswrapper[4818]: > Nov 22 04:53:35 crc kubenswrapper[4818]: I1122 04:53:35.821653 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-p7spb" podUID="1146dedf-aeff-4812-84e5-dcfe1cfc56ca" containerName="registry-server" probeResult="failure" output=< Nov 22 04:53:35 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Nov 22 04:53:35 crc kubenswrapper[4818]: > Nov 22 04:53:35 crc kubenswrapper[4818]: I1122 04:53:35.821912 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-pg659" podUID="30e236a6-d269-4f06-ae6e-e3c9bb74b4bf" containerName="registry-server" probeResult="failure" output=< Nov 22 04:53:35 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Nov 22 04:53:35 crc kubenswrapper[4818]: > Nov 22 04:53:38 crc kubenswrapper[4818]: I1122 04:53:38.438852 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9p4pn" event={"ID":"c6499ef8-b253-4710-8e30-9546db961552","Type":"ContainerStarted","Data":"0bbd8ce92cc43ddf2e192767851efa6bcf09ff583de5144e487921f42ae284dc"} Nov 22 04:53:40 crc kubenswrapper[4818]: I1122 04:53:40.477695 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9p4pn" podStartSLOduration=5.338668729 podStartE2EDuration="3m33.477664812s" podCreationTimestamp="2025-11-22 04:50:07 +0000 UTC" firstStartedPulling="2025-11-22 04:50:08.820237588 +0000 UTC m=+161.394654125" lastFinishedPulling="2025-11-22 04:53:36.959233641 +0000 UTC m=+369.533650208" observedRunningTime="2025-11-22 04:53:40.473546301 +0000 UTC m=+373.047962888" watchObservedRunningTime="2025-11-22 04:53:40.477664812 +0000 UTC m=+373.052081389" Nov 22 04:53:44 crc kubenswrapper[4818]: I1122 04:53:44.497977 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-blsms" Nov 22 04:53:44 crc kubenswrapper[4818]: I1122 04:53:44.498870 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pg659" Nov 22 04:53:44 crc kubenswrapper[4818]: I1122 04:53:44.570276 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xsv9c" Nov 22 04:53:44 crc kubenswrapper[4818]: I1122 04:53:44.581920 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pg659" Nov 22 04:53:44 crc kubenswrapper[4818]: I1122 04:53:44.598621 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-blsms" Nov 22 04:53:44 crc kubenswrapper[4818]: I1122 04:53:44.654947 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xsv9c" Nov 22 04:53:44 crc kubenswrapper[4818]: I1122 04:53:44.732508 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p7spb" Nov 22 04:53:44 crc kubenswrapper[4818]: I1122 04:53:44.783850 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p7spb" Nov 22 04:53:46 crc kubenswrapper[4818]: I1122 04:53:46.744818 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xsv9c"] Nov 22 04:53:46 crc kubenswrapper[4818]: I1122 04:53:46.745124 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xsv9c" podUID="b5d474c0-a83d-4383-a4d9-6fd0e128d34b" containerName="registry-server" containerID="cri-o://c2acf0ae27cecbf073af179865e66acb6ebcf25552d30adce84399fb8fcd2ffa" gracePeriod=2 Nov 22 04:53:46 crc kubenswrapper[4818]: I1122 04:53:46.950458 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p7spb"] Nov 22 04:53:46 crc kubenswrapper[4818]: I1122 04:53:46.950810 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p7spb" podUID="1146dedf-aeff-4812-84e5-dcfe1cfc56ca" containerName="registry-server" containerID="cri-o://66f2f515a6668df57bfe3138c5246f9ef9e6362092b0143e711b709080abe811" gracePeriod=2 Nov 22 04:53:47 crc kubenswrapper[4818]: I1122 04:53:47.635288 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9p4pn" Nov 22 04:53:47 crc kubenswrapper[4818]: I1122 04:53:47.635352 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9p4pn" Nov 22 04:53:47 crc kubenswrapper[4818]: I1122 04:53:47.703690 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9p4pn" Nov 22 04:53:48 crc kubenswrapper[4818]: I1122 04:53:48.503472 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xg9d" event={"ID":"a48b14cb-0897-4530-8f74-bee9352bcc44","Type":"ContainerStarted","Data":"6b01728a70c0148d741315caf0a86df9846c8ca4728cbb8cbfd91db0f8c6220d"} Nov 22 04:53:48 crc kubenswrapper[4818]: I1122 04:53:48.557492 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9p4pn" Nov 22 04:53:50 crc kubenswrapper[4818]: I1122 04:53:50.614784 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5n24q" event={"ID":"618de3ee-2743-44e0-9d2d-b85b1397d056","Type":"ContainerStarted","Data":"e443fdd6a4b26734c1985b58f83459a4fabc8719d5e1e443da38ea30aba334ff"} Nov 22 04:53:50 crc kubenswrapper[4818]: I1122 04:53:50.616803 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jswlc" event={"ID":"e9310d3a-0f3d-4b29-9ab6-ea9480f40274","Type":"ContainerStarted","Data":"39298509c0016a0bef2207b2fe135731a04f371951d798aec582aeebea5e3e37"} Nov 22 04:53:51 crc kubenswrapper[4818]: I1122 04:53:51.550873 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9p4pn"] Nov 22 04:53:51 crc kubenswrapper[4818]: I1122 04:53:51.551243 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9p4pn" podUID="c6499ef8-b253-4710-8e30-9546db961552" containerName="registry-server" containerID="cri-o://0bbd8ce92cc43ddf2e192767851efa6bcf09ff583de5144e487921f42ae284dc" gracePeriod=2 Nov 22 04:53:51 crc kubenswrapper[4818]: I1122 04:53:51.624916 4818 generic.go:334] "Generic (PLEG): container finished" podID="b5d474c0-a83d-4383-a4d9-6fd0e128d34b" containerID="c2acf0ae27cecbf073af179865e66acb6ebcf25552d30adce84399fb8fcd2ffa" exitCode=0 Nov 22 04:53:51 crc kubenswrapper[4818]: I1122 04:53:51.624969 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xsv9c" event={"ID":"b5d474c0-a83d-4383-a4d9-6fd0e128d34b","Type":"ContainerDied","Data":"c2acf0ae27cecbf073af179865e66acb6ebcf25552d30adce84399fb8fcd2ffa"} Nov 22 04:53:51 crc kubenswrapper[4818]: I1122 04:53:51.627347 4818 generic.go:334] "Generic (PLEG): container finished" podID="1146dedf-aeff-4812-84e5-dcfe1cfc56ca" containerID="66f2f515a6668df57bfe3138c5246f9ef9e6362092b0143e711b709080abe811" exitCode=0 Nov 22 04:53:51 crc kubenswrapper[4818]: I1122 04:53:51.627412 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7spb" event={"ID":"1146dedf-aeff-4812-84e5-dcfe1cfc56ca","Type":"ContainerDied","Data":"66f2f515a6668df57bfe3138c5246f9ef9e6362092b0143e711b709080abe811"} Nov 22 04:53:51 crc kubenswrapper[4818]: I1122 04:53:51.659363 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5n24q" podStartSLOduration=10.784386267 podStartE2EDuration="3m45.659344474s" podCreationTimestamp="2025-11-22 04:50:06 +0000 UTC" firstStartedPulling="2025-11-22 04:50:08.816921688 +0000 UTC m=+161.391338225" lastFinishedPulling="2025-11-22 04:53:43.691879865 +0000 UTC m=+376.266296432" observedRunningTime="2025-11-22 04:53:51.655147351 +0000 UTC m=+384.229563898" watchObservedRunningTime="2025-11-22 04:53:51.659344474 +0000 UTC m=+384.233761011" Nov 22 04:53:51 crc kubenswrapper[4818]: I1122 04:53:51.679515 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7xg9d" podStartSLOduration=10.785934443 podStartE2EDuration="3m45.679499136s" podCreationTimestamp="2025-11-22 04:50:06 +0000 UTC" firstStartedPulling="2025-11-22 04:50:08.799367311 +0000 UTC m=+161.373783848" lastFinishedPulling="2025-11-22 04:53:43.692931984 +0000 UTC m=+376.267348541" observedRunningTime="2025-11-22 04:53:51.673869565 +0000 UTC m=+384.248286112" watchObservedRunningTime="2025-11-22 04:53:51.679499136 +0000 UTC m=+384.253915673" Nov 22 04:53:52 crc kubenswrapper[4818]: I1122 04:53:52.670443 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jswlc" podStartSLOduration=12.514187733 podStartE2EDuration="3m47.670418068s" podCreationTimestamp="2025-11-22 04:50:05 +0000 UTC" firstStartedPulling="2025-11-22 04:50:07.778526771 +0000 UTC m=+160.352943298" lastFinishedPulling="2025-11-22 04:53:42.934757106 +0000 UTC m=+375.509173633" observedRunningTime="2025-11-22 04:53:52.667708134 +0000 UTC m=+385.242124711" watchObservedRunningTime="2025-11-22 04:53:52.670418068 +0000 UTC m=+385.244834635" Nov 22 04:53:53 crc kubenswrapper[4818]: I1122 04:53:53.851565 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xsv9c" Nov 22 04:53:53 crc kubenswrapper[4818]: I1122 04:53:53.954859 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5d474c0-a83d-4383-a4d9-6fd0e128d34b-catalog-content\") pod \"b5d474c0-a83d-4383-a4d9-6fd0e128d34b\" (UID: \"b5d474c0-a83d-4383-a4d9-6fd0e128d34b\") " Nov 22 04:53:53 crc kubenswrapper[4818]: I1122 04:53:53.954929 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lggqt\" (UniqueName: \"kubernetes.io/projected/b5d474c0-a83d-4383-a4d9-6fd0e128d34b-kube-api-access-lggqt\") pod \"b5d474c0-a83d-4383-a4d9-6fd0e128d34b\" (UID: \"b5d474c0-a83d-4383-a4d9-6fd0e128d34b\") " Nov 22 04:53:53 crc kubenswrapper[4818]: I1122 04:53:53.955000 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5d474c0-a83d-4383-a4d9-6fd0e128d34b-utilities\") pod \"b5d474c0-a83d-4383-a4d9-6fd0e128d34b\" (UID: \"b5d474c0-a83d-4383-a4d9-6fd0e128d34b\") " Nov 22 04:53:53 crc kubenswrapper[4818]: I1122 04:53:53.956966 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5d474c0-a83d-4383-a4d9-6fd0e128d34b-utilities" (OuterVolumeSpecName: "utilities") pod "b5d474c0-a83d-4383-a4d9-6fd0e128d34b" (UID: "b5d474c0-a83d-4383-a4d9-6fd0e128d34b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:53:53 crc kubenswrapper[4818]: I1122 04:53:53.964474 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5d474c0-a83d-4383-a4d9-6fd0e128d34b-kube-api-access-lggqt" (OuterVolumeSpecName: "kube-api-access-lggqt") pod "b5d474c0-a83d-4383-a4d9-6fd0e128d34b" (UID: "b5d474c0-a83d-4383-a4d9-6fd0e128d34b"). InnerVolumeSpecName "kube-api-access-lggqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.009580 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5d474c0-a83d-4383-a4d9-6fd0e128d34b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5d474c0-a83d-4383-a4d9-6fd0e128d34b" (UID: "b5d474c0-a83d-4383-a4d9-6fd0e128d34b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.057031 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5d474c0-a83d-4383-a4d9-6fd0e128d34b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.057065 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lggqt\" (UniqueName: \"kubernetes.io/projected/b5d474c0-a83d-4383-a4d9-6fd0e128d34b-kube-api-access-lggqt\") on node \"crc\" DevicePath \"\"" Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.057081 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5d474c0-a83d-4383-a4d9-6fd0e128d34b-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.186004 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p7spb" Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.261791 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkf9w\" (UniqueName: \"kubernetes.io/projected/1146dedf-aeff-4812-84e5-dcfe1cfc56ca-kube-api-access-kkf9w\") pod \"1146dedf-aeff-4812-84e5-dcfe1cfc56ca\" (UID: \"1146dedf-aeff-4812-84e5-dcfe1cfc56ca\") " Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.261899 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1146dedf-aeff-4812-84e5-dcfe1cfc56ca-utilities\") pod \"1146dedf-aeff-4812-84e5-dcfe1cfc56ca\" (UID: \"1146dedf-aeff-4812-84e5-dcfe1cfc56ca\") " Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.261972 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1146dedf-aeff-4812-84e5-dcfe1cfc56ca-catalog-content\") pod \"1146dedf-aeff-4812-84e5-dcfe1cfc56ca\" (UID: \"1146dedf-aeff-4812-84e5-dcfe1cfc56ca\") " Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.263531 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1146dedf-aeff-4812-84e5-dcfe1cfc56ca-utilities" (OuterVolumeSpecName: "utilities") pod "1146dedf-aeff-4812-84e5-dcfe1cfc56ca" (UID: "1146dedf-aeff-4812-84e5-dcfe1cfc56ca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.270570 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1146dedf-aeff-4812-84e5-dcfe1cfc56ca-kube-api-access-kkf9w" (OuterVolumeSpecName: "kube-api-access-kkf9w") pod "1146dedf-aeff-4812-84e5-dcfe1cfc56ca" (UID: "1146dedf-aeff-4812-84e5-dcfe1cfc56ca"). InnerVolumeSpecName "kube-api-access-kkf9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.319232 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1146dedf-aeff-4812-84e5-dcfe1cfc56ca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1146dedf-aeff-4812-84e5-dcfe1cfc56ca" (UID: "1146dedf-aeff-4812-84e5-dcfe1cfc56ca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.364489 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkf9w\" (UniqueName: \"kubernetes.io/projected/1146dedf-aeff-4812-84e5-dcfe1cfc56ca-kube-api-access-kkf9w\") on node \"crc\" DevicePath \"\"" Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.364884 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1146dedf-aeff-4812-84e5-dcfe1cfc56ca-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.364901 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1146dedf-aeff-4812-84e5-dcfe1cfc56ca-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.667825 4818 generic.go:334] "Generic (PLEG): container finished" podID="c6499ef8-b253-4710-8e30-9546db961552" containerID="0bbd8ce92cc43ddf2e192767851efa6bcf09ff583de5144e487921f42ae284dc" exitCode=0 Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.667938 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9p4pn" event={"ID":"c6499ef8-b253-4710-8e30-9546db961552","Type":"ContainerDied","Data":"0bbd8ce92cc43ddf2e192767851efa6bcf09ff583de5144e487921f42ae284dc"} Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.671550 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7spb" event={"ID":"1146dedf-aeff-4812-84e5-dcfe1cfc56ca","Type":"ContainerDied","Data":"cfcd480d80a31ea0921b3db61192d1a184c3c326afcb733df013a4fbd39fa753"} Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.671608 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p7spb" Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.671627 4818 scope.go:117] "RemoveContainer" containerID="66f2f515a6668df57bfe3138c5246f9ef9e6362092b0143e711b709080abe811" Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.674781 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xsv9c" event={"ID":"b5d474c0-a83d-4383-a4d9-6fd0e128d34b","Type":"ContainerDied","Data":"8f7cfcbd668cf10f2fa4d8b296d4584a206d1c4c5b32cb57e33f37e02cad47ee"} Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.675108 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xsv9c" Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.705174 4818 scope.go:117] "RemoveContainer" containerID="3a15bf524b1d9f4914e4ad0b1a97fce97e8e952dc7c7bb977cd258cff18dd7c3" Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.715512 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xsv9c"] Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.724968 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xsv9c"] Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.733934 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p7spb"] Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.734102 4818 scope.go:117] "RemoveContainer" containerID="4104ec4e5e0aa69fa66743a91c89553a4bdfff3e7c692859b10f413333cadc3e" Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.739785 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p7spb"] Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.758859 4818 scope.go:117] "RemoveContainer" containerID="c2acf0ae27cecbf073af179865e66acb6ebcf25552d30adce84399fb8fcd2ffa" Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.782553 4818 scope.go:117] "RemoveContainer" containerID="8a90bc14671a07eadef9a712e2c4e9b9276f9543f2512b4f68c64b6bb097c014" Nov 22 04:53:54 crc kubenswrapper[4818]: I1122 04:53:54.802382 4818 scope.go:117] "RemoveContainer" containerID="8499a55582668439375b5171b8bdb3172c405b079e6e0d74af514f3f7d54b04c" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.180665 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9p4pn" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.238079 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jswlc" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.238508 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jswlc" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.291804 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7srn\" (UniqueName: \"kubernetes.io/projected/c6499ef8-b253-4710-8e30-9546db961552-kube-api-access-n7srn\") pod \"c6499ef8-b253-4710-8e30-9546db961552\" (UID: \"c6499ef8-b253-4710-8e30-9546db961552\") " Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.292222 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6499ef8-b253-4710-8e30-9546db961552-utilities\") pod \"c6499ef8-b253-4710-8e30-9546db961552\" (UID: \"c6499ef8-b253-4710-8e30-9546db961552\") " Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.292345 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6499ef8-b253-4710-8e30-9546db961552-catalog-content\") pod \"c6499ef8-b253-4710-8e30-9546db961552\" (UID: \"c6499ef8-b253-4710-8e30-9546db961552\") " Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.293719 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6499ef8-b253-4710-8e30-9546db961552-utilities" (OuterVolumeSpecName: "utilities") pod "c6499ef8-b253-4710-8e30-9546db961552" (UID: "c6499ef8-b253-4710-8e30-9546db961552"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.300511 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6499ef8-b253-4710-8e30-9546db961552-kube-api-access-n7srn" (OuterVolumeSpecName: "kube-api-access-n7srn") pod "c6499ef8-b253-4710-8e30-9546db961552" (UID: "c6499ef8-b253-4710-8e30-9546db961552"). InnerVolumeSpecName "kube-api-access-n7srn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.304642 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1146dedf-aeff-4812-84e5-dcfe1cfc56ca" path="/var/lib/kubelet/pods/1146dedf-aeff-4812-84e5-dcfe1cfc56ca/volumes" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.305928 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5d474c0-a83d-4383-a4d9-6fd0e128d34b" path="/var/lib/kubelet/pods/b5d474c0-a83d-4383-a4d9-6fd0e128d34b/volumes" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.337221 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jswlc" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.394999 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7srn\" (UniqueName: \"kubernetes.io/projected/c6499ef8-b253-4710-8e30-9546db961552-kube-api-access-n7srn\") on node \"crc\" DevicePath \"\"" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.395107 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6499ef8-b253-4710-8e30-9546db961552-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.587929 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6499ef8-b253-4710-8e30-9546db961552-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c6499ef8-b253-4710-8e30-9546db961552" (UID: "c6499ef8-b253-4710-8e30-9546db961552"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.597808 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6499ef8-b253-4710-8e30-9546db961552-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.650035 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5n24q" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.650443 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5n24q" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.696684 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9p4pn" event={"ID":"c6499ef8-b253-4710-8e30-9546db961552","Type":"ContainerDied","Data":"2d806f2c5c3db0b2ff4894b07736e27b0aaf0fd425d80773189028ed97a31874"} Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.696777 4818 scope.go:117] "RemoveContainer" containerID="0bbd8ce92cc43ddf2e192767851efa6bcf09ff583de5144e487921f42ae284dc" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.697018 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9p4pn" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.725034 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5n24q" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.725349 4818 scope.go:117] "RemoveContainer" containerID="0d40aadf30061933224d26c5bc9dfd8641cd22b347bd83e655cfde85f5d25412" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.762421 4818 scope.go:117] "RemoveContainer" containerID="7bd756d1b02621137e66144ce97af945367c246176629efd9f82e207a32441ee" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.772171 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jswlc" Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.864241 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9p4pn"] Nov 22 04:53:56 crc kubenswrapper[4818]: I1122 04:53:56.867967 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9p4pn"] Nov 22 04:53:57 crc kubenswrapper[4818]: I1122 04:53:57.268207 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7xg9d" Nov 22 04:53:57 crc kubenswrapper[4818]: I1122 04:53:57.268281 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7xg9d" Nov 22 04:53:57 crc kubenswrapper[4818]: I1122 04:53:57.329715 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7xg9d" Nov 22 04:53:57 crc kubenswrapper[4818]: I1122 04:53:57.774446 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5n24q" Nov 22 04:53:57 crc kubenswrapper[4818]: I1122 04:53:57.776218 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7xg9d" Nov 22 04:53:58 crc kubenswrapper[4818]: I1122 04:53:58.303201 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6499ef8-b253-4710-8e30-9546db961552" path="/var/lib/kubelet/pods/c6499ef8-b253-4710-8e30-9546db961552/volumes" Nov 22 04:54:00 crc kubenswrapper[4818]: I1122 04:54:00.343487 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5n24q"] Nov 22 04:54:00 crc kubenswrapper[4818]: I1122 04:54:00.343753 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5n24q" podUID="618de3ee-2743-44e0-9d2d-b85b1397d056" containerName="registry-server" containerID="cri-o://e443fdd6a4b26734c1985b58f83459a4fabc8719d5e1e443da38ea30aba334ff" gracePeriod=2 Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.604122 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5n24q" Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.677183 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/618de3ee-2743-44e0-9d2d-b85b1397d056-catalog-content\") pod \"618de3ee-2743-44e0-9d2d-b85b1397d056\" (UID: \"618de3ee-2743-44e0-9d2d-b85b1397d056\") " Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.677335 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/618de3ee-2743-44e0-9d2d-b85b1397d056-utilities\") pod \"618de3ee-2743-44e0-9d2d-b85b1397d056\" (UID: \"618de3ee-2743-44e0-9d2d-b85b1397d056\") " Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.677490 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p26q4\" (UniqueName: \"kubernetes.io/projected/618de3ee-2743-44e0-9d2d-b85b1397d056-kube-api-access-p26q4\") pod \"618de3ee-2743-44e0-9d2d-b85b1397d056\" (UID: \"618de3ee-2743-44e0-9d2d-b85b1397d056\") " Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.678435 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/618de3ee-2743-44e0-9d2d-b85b1397d056-utilities" (OuterVolumeSpecName: "utilities") pod "618de3ee-2743-44e0-9d2d-b85b1397d056" (UID: "618de3ee-2743-44e0-9d2d-b85b1397d056"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.687473 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/618de3ee-2743-44e0-9d2d-b85b1397d056-kube-api-access-p26q4" (OuterVolumeSpecName: "kube-api-access-p26q4") pod "618de3ee-2743-44e0-9d2d-b85b1397d056" (UID: "618de3ee-2743-44e0-9d2d-b85b1397d056"). InnerVolumeSpecName "kube-api-access-p26q4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.695536 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/618de3ee-2743-44e0-9d2d-b85b1397d056-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "618de3ee-2743-44e0-9d2d-b85b1397d056" (UID: "618de3ee-2743-44e0-9d2d-b85b1397d056"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.734039 4818 generic.go:334] "Generic (PLEG): container finished" podID="618de3ee-2743-44e0-9d2d-b85b1397d056" containerID="e443fdd6a4b26734c1985b58f83459a4fabc8719d5e1e443da38ea30aba334ff" exitCode=0 Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.734087 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5n24q" event={"ID":"618de3ee-2743-44e0-9d2d-b85b1397d056","Type":"ContainerDied","Data":"e443fdd6a4b26734c1985b58f83459a4fabc8719d5e1e443da38ea30aba334ff"} Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.734114 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5n24q" event={"ID":"618de3ee-2743-44e0-9d2d-b85b1397d056","Type":"ContainerDied","Data":"8b7542a2f23c54b8ef16cac0ed6dd7ff8cd9f382f83ac5f7032c2d3c026b9caf"} Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.734134 4818 scope.go:117] "RemoveContainer" containerID="e443fdd6a4b26734c1985b58f83459a4fabc8719d5e1e443da38ea30aba334ff" Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.734307 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5n24q" Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.754467 4818 scope.go:117] "RemoveContainer" containerID="2ee78d0334a79dce9184e6181b232555a3e2863948649a205067236c72d4ed84" Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.771004 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5n24q"] Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.773782 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5n24q"] Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.778644 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p26q4\" (UniqueName: \"kubernetes.io/projected/618de3ee-2743-44e0-9d2d-b85b1397d056-kube-api-access-p26q4\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.778672 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/618de3ee-2743-44e0-9d2d-b85b1397d056-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.778686 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/618de3ee-2743-44e0-9d2d-b85b1397d056-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.788488 4818 scope.go:117] "RemoveContainer" containerID="e131704fbc499124d84a87e92561df99983f1b2b6e26d37ae1856073fca84532" Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.805414 4818 scope.go:117] "RemoveContainer" containerID="e443fdd6a4b26734c1985b58f83459a4fabc8719d5e1e443da38ea30aba334ff" Nov 22 04:54:01 crc kubenswrapper[4818]: E1122 04:54:01.805888 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e443fdd6a4b26734c1985b58f83459a4fabc8719d5e1e443da38ea30aba334ff\": container with ID starting with e443fdd6a4b26734c1985b58f83459a4fabc8719d5e1e443da38ea30aba334ff not found: ID does not exist" containerID="e443fdd6a4b26734c1985b58f83459a4fabc8719d5e1e443da38ea30aba334ff" Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.805950 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e443fdd6a4b26734c1985b58f83459a4fabc8719d5e1e443da38ea30aba334ff"} err="failed to get container status \"e443fdd6a4b26734c1985b58f83459a4fabc8719d5e1e443da38ea30aba334ff\": rpc error: code = NotFound desc = could not find container \"e443fdd6a4b26734c1985b58f83459a4fabc8719d5e1e443da38ea30aba334ff\": container with ID starting with e443fdd6a4b26734c1985b58f83459a4fabc8719d5e1e443da38ea30aba334ff not found: ID does not exist" Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.805986 4818 scope.go:117] "RemoveContainer" containerID="2ee78d0334a79dce9184e6181b232555a3e2863948649a205067236c72d4ed84" Nov 22 04:54:01 crc kubenswrapper[4818]: E1122 04:54:01.806432 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ee78d0334a79dce9184e6181b232555a3e2863948649a205067236c72d4ed84\": container with ID starting with 2ee78d0334a79dce9184e6181b232555a3e2863948649a205067236c72d4ed84 not found: ID does not exist" containerID="2ee78d0334a79dce9184e6181b232555a3e2863948649a205067236c72d4ed84" Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.806478 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ee78d0334a79dce9184e6181b232555a3e2863948649a205067236c72d4ed84"} err="failed to get container status \"2ee78d0334a79dce9184e6181b232555a3e2863948649a205067236c72d4ed84\": rpc error: code = NotFound desc = could not find container \"2ee78d0334a79dce9184e6181b232555a3e2863948649a205067236c72d4ed84\": container with ID starting with 2ee78d0334a79dce9184e6181b232555a3e2863948649a205067236c72d4ed84 not found: ID does not exist" Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.806510 4818 scope.go:117] "RemoveContainer" containerID="e131704fbc499124d84a87e92561df99983f1b2b6e26d37ae1856073fca84532" Nov 22 04:54:01 crc kubenswrapper[4818]: E1122 04:54:01.806844 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e131704fbc499124d84a87e92561df99983f1b2b6e26d37ae1856073fca84532\": container with ID starting with e131704fbc499124d84a87e92561df99983f1b2b6e26d37ae1856073fca84532 not found: ID does not exist" containerID="e131704fbc499124d84a87e92561df99983f1b2b6e26d37ae1856073fca84532" Nov 22 04:54:01 crc kubenswrapper[4818]: I1122 04:54:01.806876 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e131704fbc499124d84a87e92561df99983f1b2b6e26d37ae1856073fca84532"} err="failed to get container status \"e131704fbc499124d84a87e92561df99983f1b2b6e26d37ae1856073fca84532\": rpc error: code = NotFound desc = could not find container \"e131704fbc499124d84a87e92561df99983f1b2b6e26d37ae1856073fca84532\": container with ID starting with e131704fbc499124d84a87e92561df99983f1b2b6e26d37ae1856073fca84532 not found: ID does not exist" Nov 22 04:54:02 crc kubenswrapper[4818]: I1122 04:54:02.310377 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="618de3ee-2743-44e0-9d2d-b85b1397d056" path="/var/lib/kubelet/pods/618de3ee-2743-44e0-9d2d-b85b1397d056/volumes" Nov 22 04:54:09 crc kubenswrapper[4818]: I1122 04:54:09.681361 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mkgj4"] Nov 22 04:54:21 crc kubenswrapper[4818]: I1122 04:54:21.264932 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 04:54:21 crc kubenswrapper[4818]: I1122 04:54:21.265646 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 04:54:34 crc kubenswrapper[4818]: I1122 04:54:34.709410 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" podUID="0dcaa80a-8993-4452-9acb-b8010eade7c6" containerName="oauth-openshift" containerID="cri-o://7ac0f02857258c046abc9f0e943dc57129e4b404eb59e0b7b31e34bc54c99a9c" gracePeriod=15 Nov 22 04:54:35 crc kubenswrapper[4818]: I1122 04:54:35.976178 4818 generic.go:334] "Generic (PLEG): container finished" podID="0dcaa80a-8993-4452-9acb-b8010eade7c6" containerID="7ac0f02857258c046abc9f0e943dc57129e4b404eb59e0b7b31e34bc54c99a9c" exitCode=0 Nov 22 04:54:35 crc kubenswrapper[4818]: I1122 04:54:35.976307 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" event={"ID":"0dcaa80a-8993-4452-9acb-b8010eade7c6","Type":"ContainerDied","Data":"7ac0f02857258c046abc9f0e943dc57129e4b404eb59e0b7b31e34bc54c99a9c"} Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.438746 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.476755 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv"] Nov 22 04:54:36 crc kubenswrapper[4818]: E1122 04:54:36.477102 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b4aabc4-c336-4114-86f4-4f94bb336609" containerName="pruner" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.477149 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b4aabc4-c336-4114-86f4-4f94bb336609" containerName="pruner" Nov 22 04:54:36 crc kubenswrapper[4818]: E1122 04:54:36.477163 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6499ef8-b253-4710-8e30-9546db961552" containerName="extract-utilities" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.477171 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6499ef8-b253-4710-8e30-9546db961552" containerName="extract-utilities" Nov 22 04:54:36 crc kubenswrapper[4818]: E1122 04:54:36.477183 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1146dedf-aeff-4812-84e5-dcfe1cfc56ca" containerName="extract-content" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.477219 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="1146dedf-aeff-4812-84e5-dcfe1cfc56ca" containerName="extract-content" Nov 22 04:54:36 crc kubenswrapper[4818]: E1122 04:54:36.477233 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1146dedf-aeff-4812-84e5-dcfe1cfc56ca" containerName="extract-utilities" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.477243 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="1146dedf-aeff-4812-84e5-dcfe1cfc56ca" containerName="extract-utilities" Nov 22 04:54:36 crc kubenswrapper[4818]: E1122 04:54:36.477294 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1146dedf-aeff-4812-84e5-dcfe1cfc56ca" containerName="registry-server" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.477302 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="1146dedf-aeff-4812-84e5-dcfe1cfc56ca" containerName="registry-server" Nov 22 04:54:36 crc kubenswrapper[4818]: E1122 04:54:36.477312 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5d474c0-a83d-4383-a4d9-6fd0e128d34b" containerName="registry-server" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.477321 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d474c0-a83d-4383-a4d9-6fd0e128d34b" containerName="registry-server" Nov 22 04:54:36 crc kubenswrapper[4818]: E1122 04:54:36.477336 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="618de3ee-2743-44e0-9d2d-b85b1397d056" containerName="extract-utilities" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.477373 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="618de3ee-2743-44e0-9d2d-b85b1397d056" containerName="extract-utilities" Nov 22 04:54:36 crc kubenswrapper[4818]: E1122 04:54:36.477390 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="618de3ee-2743-44e0-9d2d-b85b1397d056" containerName="extract-content" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.477398 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="618de3ee-2743-44e0-9d2d-b85b1397d056" containerName="extract-content" Nov 22 04:54:36 crc kubenswrapper[4818]: E1122 04:54:36.477411 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5d474c0-a83d-4383-a4d9-6fd0e128d34b" containerName="extract-content" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.477419 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d474c0-a83d-4383-a4d9-6fd0e128d34b" containerName="extract-content" Nov 22 04:54:36 crc kubenswrapper[4818]: E1122 04:54:36.477458 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6499ef8-b253-4710-8e30-9546db961552" containerName="registry-server" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.477466 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6499ef8-b253-4710-8e30-9546db961552" containerName="registry-server" Nov 22 04:54:36 crc kubenswrapper[4818]: E1122 04:54:36.477477 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="618de3ee-2743-44e0-9d2d-b85b1397d056" containerName="registry-server" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.477485 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="618de3ee-2743-44e0-9d2d-b85b1397d056" containerName="registry-server" Nov 22 04:54:36 crc kubenswrapper[4818]: E1122 04:54:36.477500 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5d474c0-a83d-4383-a4d9-6fd0e128d34b" containerName="extract-utilities" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.477533 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d474c0-a83d-4383-a4d9-6fd0e128d34b" containerName="extract-utilities" Nov 22 04:54:36 crc kubenswrapper[4818]: E1122 04:54:36.477547 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6499ef8-b253-4710-8e30-9546db961552" containerName="extract-content" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.477555 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6499ef8-b253-4710-8e30-9546db961552" containerName="extract-content" Nov 22 04:54:36 crc kubenswrapper[4818]: E1122 04:54:36.477564 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dcaa80a-8993-4452-9acb-b8010eade7c6" containerName="oauth-openshift" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.477572 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dcaa80a-8993-4452-9acb-b8010eade7c6" containerName="oauth-openshift" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.477731 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b4aabc4-c336-4114-86f4-4f94bb336609" containerName="pruner" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.477744 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6499ef8-b253-4710-8e30-9546db961552" containerName="registry-server" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.477780 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="618de3ee-2743-44e0-9d2d-b85b1397d056" containerName="registry-server" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.477792 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5d474c0-a83d-4383-a4d9-6fd0e128d34b" containerName="registry-server" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.477827 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="1146dedf-aeff-4812-84e5-dcfe1cfc56ca" containerName="registry-server" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.477882 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dcaa80a-8993-4452-9acb-b8010eade7c6" containerName="oauth-openshift" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.478461 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.500793 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv"] Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.521721 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-idp-0-file-data\") pod \"0dcaa80a-8993-4452-9acb-b8010eade7c6\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.521822 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-template-provider-selection\") pod \"0dcaa80a-8993-4452-9acb-b8010eade7c6\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.521877 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-serving-cert\") pod \"0dcaa80a-8993-4452-9acb-b8010eade7c6\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.521962 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-template-error\") pod \"0dcaa80a-8993-4452-9acb-b8010eade7c6\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.522014 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-cliconfig\") pod \"0dcaa80a-8993-4452-9acb-b8010eade7c6\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.522056 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-trusted-ca-bundle\") pod \"0dcaa80a-8993-4452-9acb-b8010eade7c6\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.522100 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjkhp\" (UniqueName: \"kubernetes.io/projected/0dcaa80a-8993-4452-9acb-b8010eade7c6-kube-api-access-sjkhp\") pod \"0dcaa80a-8993-4452-9acb-b8010eade7c6\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.522146 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-service-ca\") pod \"0dcaa80a-8993-4452-9acb-b8010eade7c6\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.522180 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-audit-policies\") pod \"0dcaa80a-8993-4452-9acb-b8010eade7c6\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.522211 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-template-login\") pod \"0dcaa80a-8993-4452-9acb-b8010eade7c6\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.522240 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-session\") pod \"0dcaa80a-8993-4452-9acb-b8010eade7c6\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.522296 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0dcaa80a-8993-4452-9acb-b8010eade7c6-audit-dir\") pod \"0dcaa80a-8993-4452-9acb-b8010eade7c6\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.522363 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-ocp-branding-template\") pod \"0dcaa80a-8993-4452-9acb-b8010eade7c6\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.522410 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-router-certs\") pod \"0dcaa80a-8993-4452-9acb-b8010eade7c6\" (UID: \"0dcaa80a-8993-4452-9acb-b8010eade7c6\") " Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.522647 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-router-certs\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.522693 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.522728 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-user-template-login\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.522765 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-service-ca\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.522810 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e0e6a106-e521-4000-890b-f1cc57e0176b-audit-dir\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.522803 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "0dcaa80a-8993-4452-9acb-b8010eade7c6" (UID: "0dcaa80a-8993-4452-9acb-b8010eade7c6"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.522851 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dd7f\" (UniqueName: \"kubernetes.io/projected/e0e6a106-e521-4000-890b-f1cc57e0176b-kube-api-access-8dd7f\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.522837 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "0dcaa80a-8993-4452-9acb-b8010eade7c6" (UID: "0dcaa80a-8993-4452-9acb-b8010eade7c6"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.522985 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.523018 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.523073 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-session\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.523127 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.523170 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e0e6a106-e521-4000-890b-f1cc57e0176b-audit-policies\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.523199 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-user-template-error\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.523230 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.524201 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.524311 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.524341 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.523549 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "0dcaa80a-8993-4452-9acb-b8010eade7c6" (UID: "0dcaa80a-8993-4452-9acb-b8010eade7c6"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.524019 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "0dcaa80a-8993-4452-9acb-b8010eade7c6" (UID: "0dcaa80a-8993-4452-9acb-b8010eade7c6"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.524064 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dcaa80a-8993-4452-9acb-b8010eade7c6-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "0dcaa80a-8993-4452-9acb-b8010eade7c6" (UID: "0dcaa80a-8993-4452-9acb-b8010eade7c6"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.528914 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "0dcaa80a-8993-4452-9acb-b8010eade7c6" (UID: "0dcaa80a-8993-4452-9acb-b8010eade7c6"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.529547 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "0dcaa80a-8993-4452-9acb-b8010eade7c6" (UID: "0dcaa80a-8993-4452-9acb-b8010eade7c6"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.533125 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dcaa80a-8993-4452-9acb-b8010eade7c6-kube-api-access-sjkhp" (OuterVolumeSpecName: "kube-api-access-sjkhp") pod "0dcaa80a-8993-4452-9acb-b8010eade7c6" (UID: "0dcaa80a-8993-4452-9acb-b8010eade7c6"). InnerVolumeSpecName "kube-api-access-sjkhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.533862 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "0dcaa80a-8993-4452-9acb-b8010eade7c6" (UID: "0dcaa80a-8993-4452-9acb-b8010eade7c6"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.534937 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "0dcaa80a-8993-4452-9acb-b8010eade7c6" (UID: "0dcaa80a-8993-4452-9acb-b8010eade7c6"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.535471 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "0dcaa80a-8993-4452-9acb-b8010eade7c6" (UID: "0dcaa80a-8993-4452-9acb-b8010eade7c6"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.539922 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "0dcaa80a-8993-4452-9acb-b8010eade7c6" (UID: "0dcaa80a-8993-4452-9acb-b8010eade7c6"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.542848 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "0dcaa80a-8993-4452-9acb-b8010eade7c6" (UID: "0dcaa80a-8993-4452-9acb-b8010eade7c6"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.544841 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "0dcaa80a-8993-4452-9acb-b8010eade7c6" (UID: "0dcaa80a-8993-4452-9acb-b8010eade7c6"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626193 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626322 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e0e6a106-e521-4000-890b-f1cc57e0176b-audit-policies\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626359 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626394 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-user-template-error\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626429 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626483 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-router-certs\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626513 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626546 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-user-template-login\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626582 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-service-ca\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626630 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e0e6a106-e521-4000-890b-f1cc57e0176b-audit-dir\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626673 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dd7f\" (UniqueName: \"kubernetes.io/projected/e0e6a106-e521-4000-890b-f1cc57e0176b-kube-api-access-8dd7f\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626718 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626751 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626801 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-session\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626876 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626898 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626920 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjkhp\" (UniqueName: \"kubernetes.io/projected/0dcaa80a-8993-4452-9acb-b8010eade7c6-kube-api-access-sjkhp\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626940 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626959 4818 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0dcaa80a-8993-4452-9acb-b8010eade7c6-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626976 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.626994 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.627014 4818 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0dcaa80a-8993-4452-9acb-b8010eade7c6-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.627032 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.627051 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.627069 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.627088 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0dcaa80a-8993-4452-9acb-b8010eade7c6-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.627585 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e0e6a106-e521-4000-890b-f1cc57e0176b-audit-dir\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.629128 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-service-ca\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.629327 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.630507 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e0e6a106-e521-4000-890b-f1cc57e0176b-audit-policies\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.630644 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.633365 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.633439 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-user-template-error\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.633460 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-router-certs\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.634321 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-user-template-login\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.634354 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.635484 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-session\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.636001 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.636483 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e0e6a106-e521-4000-890b-f1cc57e0176b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.658213 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dd7f\" (UniqueName: \"kubernetes.io/projected/e0e6a106-e521-4000-890b-f1cc57e0176b-kube-api-access-8dd7f\") pod \"oauth-openshift-7c65b78cb7-bs4kv\" (UID: \"e0e6a106-e521-4000-890b-f1cc57e0176b\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.815554 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.985114 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" event={"ID":"0dcaa80a-8993-4452-9acb-b8010eade7c6","Type":"ContainerDied","Data":"2e962e5be8a55f03f46497d8f19ecb27be2fa251de108aaddc747db3eaec4fc0"} Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.985172 4818 scope.go:117] "RemoveContainer" containerID="7ac0f02857258c046abc9f0e943dc57129e4b404eb59e0b7b31e34bc54c99a9c" Nov 22 04:54:36 crc kubenswrapper[4818]: I1122 04:54:36.985334 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mkgj4" Nov 22 04:54:37 crc kubenswrapper[4818]: I1122 04:54:37.016684 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mkgj4"] Nov 22 04:54:37 crc kubenswrapper[4818]: I1122 04:54:37.029337 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mkgj4"] Nov 22 04:54:37 crc kubenswrapper[4818]: I1122 04:54:37.032884 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv"] Nov 22 04:54:37 crc kubenswrapper[4818]: W1122 04:54:37.036488 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0e6a106_e521_4000_890b_f1cc57e0176b.slice/crio-f4b24762ada23c241c18911a1b2dbfc8fb8170a16c1b499f5eecbf77bce70900 WatchSource:0}: Error finding container f4b24762ada23c241c18911a1b2dbfc8fb8170a16c1b499f5eecbf77bce70900: Status 404 returned error can't find the container with id f4b24762ada23c241c18911a1b2dbfc8fb8170a16c1b499f5eecbf77bce70900 Nov 22 04:54:38 crc kubenswrapper[4818]: I1122 04:54:37.997868 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" event={"ID":"e0e6a106-e521-4000-890b-f1cc57e0176b","Type":"ContainerStarted","Data":"1e0b183882b74b8bc155ada322e6f54f9830334c1a39a9dfb205a1e9d11dd825"} Nov 22 04:54:38 crc kubenswrapper[4818]: I1122 04:54:38.000893 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" event={"ID":"e0e6a106-e521-4000-890b-f1cc57e0176b","Type":"ContainerStarted","Data":"f4b24762ada23c241c18911a1b2dbfc8fb8170a16c1b499f5eecbf77bce70900"} Nov 22 04:54:38 crc kubenswrapper[4818]: I1122 04:54:38.000952 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:38 crc kubenswrapper[4818]: I1122 04:54:38.008205 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" Nov 22 04:54:38 crc kubenswrapper[4818]: I1122 04:54:38.020333 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7c65b78cb7-bs4kv" podStartSLOduration=29.020315873 podStartE2EDuration="29.020315873s" podCreationTimestamp="2025-11-22 04:54:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:54:38.017323723 +0000 UTC m=+430.591740280" watchObservedRunningTime="2025-11-22 04:54:38.020315873 +0000 UTC m=+430.594732420" Nov 22 04:54:38 crc kubenswrapper[4818]: I1122 04:54:38.296804 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dcaa80a-8993-4452-9acb-b8010eade7c6" path="/var/lib/kubelet/pods/0dcaa80a-8993-4452-9acb-b8010eade7c6/volumes" Nov 22 04:54:51 crc kubenswrapper[4818]: I1122 04:54:51.265218 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 04:54:51 crc kubenswrapper[4818]: I1122 04:54:51.265749 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 04:54:56 crc kubenswrapper[4818]: I1122 04:54:56.794106 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-blsms"] Nov 22 04:54:56 crc kubenswrapper[4818]: I1122 04:54:56.794710 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-blsms" podUID="0e0614cc-57b5-4782-b462-e693c03c2117" containerName="registry-server" containerID="cri-o://d57a8171c3e97f6c8a2297c4ba361e5d01f7fa77f4fbcb6d72a3d2428031fa78" gracePeriod=30 Nov 22 04:54:56 crc kubenswrapper[4818]: I1122 04:54:56.803327 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pg659"] Nov 22 04:54:56 crc kubenswrapper[4818]: I1122 04:54:56.803544 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pg659" podUID="30e236a6-d269-4f06-ae6e-e3c9bb74b4bf" containerName="registry-server" containerID="cri-o://b00ce629adadad84a158329e4f6d5007a908430dbc8e874a6cc60a17b73dbdc2" gracePeriod=30 Nov 22 04:54:56 crc kubenswrapper[4818]: I1122 04:54:56.810987 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nz99w"] Nov 22 04:54:56 crc kubenswrapper[4818]: I1122 04:54:56.811745 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" podUID="1d4ef859-1409-451a-a2e1-4c1c8c6d473b" containerName="marketplace-operator" containerID="cri-o://1d9c4f4c749f569d569ffc735c3e3244d269e0b217dcbf2dafad509d813c3355" gracePeriod=30 Nov 22 04:54:56 crc kubenswrapper[4818]: I1122 04:54:56.820527 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jswlc"] Nov 22 04:54:56 crc kubenswrapper[4818]: I1122 04:54:56.820793 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jswlc" podUID="e9310d3a-0f3d-4b29-9ab6-ea9480f40274" containerName="registry-server" containerID="cri-o://39298509c0016a0bef2207b2fe135731a04f371951d798aec582aeebea5e3e37" gracePeriod=30 Nov 22 04:54:56 crc kubenswrapper[4818]: I1122 04:54:56.833566 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cxqxn"] Nov 22 04:54:56 crc kubenswrapper[4818]: I1122 04:54:56.850204 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7xg9d"] Nov 22 04:54:56 crc kubenswrapper[4818]: I1122 04:54:56.850444 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cxqxn" Nov 22 04:54:56 crc kubenswrapper[4818]: I1122 04:54:56.850597 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7xg9d" podUID="a48b14cb-0897-4530-8f74-bee9352bcc44" containerName="registry-server" containerID="cri-o://6b01728a70c0148d741315caf0a86df9846c8ca4728cbb8cbfd91db0f8c6220d" gracePeriod=30 Nov 22 04:54:56 crc kubenswrapper[4818]: I1122 04:54:56.857696 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cxqxn"] Nov 22 04:54:56 crc kubenswrapper[4818]: I1122 04:54:56.896962 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwvhb\" (UniqueName: \"kubernetes.io/projected/6e9cbedb-a168-495f-8746-ccf21c4e330c-kube-api-access-gwvhb\") pod \"marketplace-operator-79b997595-cxqxn\" (UID: \"6e9cbedb-a168-495f-8746-ccf21c4e330c\") " pod="openshift-marketplace/marketplace-operator-79b997595-cxqxn" Nov 22 04:54:56 crc kubenswrapper[4818]: I1122 04:54:56.897048 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6e9cbedb-a168-495f-8746-ccf21c4e330c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cxqxn\" (UID: \"6e9cbedb-a168-495f-8746-ccf21c4e330c\") " pod="openshift-marketplace/marketplace-operator-79b997595-cxqxn" Nov 22 04:54:56 crc kubenswrapper[4818]: I1122 04:54:56.897085 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6e9cbedb-a168-495f-8746-ccf21c4e330c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cxqxn\" (UID: \"6e9cbedb-a168-495f-8746-ccf21c4e330c\") " pod="openshift-marketplace/marketplace-operator-79b997595-cxqxn" Nov 22 04:54:56 crc kubenswrapper[4818]: I1122 04:54:56.998178 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwvhb\" (UniqueName: \"kubernetes.io/projected/6e9cbedb-a168-495f-8746-ccf21c4e330c-kube-api-access-gwvhb\") pod \"marketplace-operator-79b997595-cxqxn\" (UID: \"6e9cbedb-a168-495f-8746-ccf21c4e330c\") " pod="openshift-marketplace/marketplace-operator-79b997595-cxqxn" Nov 22 04:54:56 crc kubenswrapper[4818]: I1122 04:54:56.998352 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6e9cbedb-a168-495f-8746-ccf21c4e330c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cxqxn\" (UID: \"6e9cbedb-a168-495f-8746-ccf21c4e330c\") " pod="openshift-marketplace/marketplace-operator-79b997595-cxqxn" Nov 22 04:54:56 crc kubenswrapper[4818]: I1122 04:54:56.998426 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6e9cbedb-a168-495f-8746-ccf21c4e330c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cxqxn\" (UID: \"6e9cbedb-a168-495f-8746-ccf21c4e330c\") " pod="openshift-marketplace/marketplace-operator-79b997595-cxqxn" Nov 22 04:54:57 crc kubenswrapper[4818]: I1122 04:54:57.125035 4818 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-nz99w container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Nov 22 04:54:57 crc kubenswrapper[4818]: I1122 04:54:57.125419 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" podUID="1d4ef859-1409-451a-a2e1-4c1c8c6d473b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Nov 22 04:54:57 crc kubenswrapper[4818]: E1122 04:54:57.269506 4818 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6b01728a70c0148d741315caf0a86df9846c8ca4728cbb8cbfd91db0f8c6220d" cmd=["grpc_health_probe","-addr=:50051"] Nov 22 04:54:57 crc kubenswrapper[4818]: E1122 04:54:57.270560 4818 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6b01728a70c0148d741315caf0a86df9846c8ca4728cbb8cbfd91db0f8c6220d" cmd=["grpc_health_probe","-addr=:50051"] Nov 22 04:54:57 crc kubenswrapper[4818]: E1122 04:54:57.271974 4818 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6b01728a70c0148d741315caf0a86df9846c8ca4728cbb8cbfd91db0f8c6220d" cmd=["grpc_health_probe","-addr=:50051"] Nov 22 04:54:57 crc kubenswrapper[4818]: E1122 04:54:57.272010 4818 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-marketplace/redhat-operators-7xg9d" podUID="a48b14cb-0897-4530-8f74-bee9352bcc44" containerName="registry-server" Nov 22 04:54:57 crc kubenswrapper[4818]: I1122 04:54:57.513272 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6e9cbedb-a168-495f-8746-ccf21c4e330c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cxqxn\" (UID: \"6e9cbedb-a168-495f-8746-ccf21c4e330c\") " pod="openshift-marketplace/marketplace-operator-79b997595-cxqxn" Nov 22 04:54:57 crc kubenswrapper[4818]: I1122 04:54:57.513439 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwvhb\" (UniqueName: \"kubernetes.io/projected/6e9cbedb-a168-495f-8746-ccf21c4e330c-kube-api-access-gwvhb\") pod \"marketplace-operator-79b997595-cxqxn\" (UID: \"6e9cbedb-a168-495f-8746-ccf21c4e330c\") " pod="openshift-marketplace/marketplace-operator-79b997595-cxqxn" Nov 22 04:54:57 crc kubenswrapper[4818]: I1122 04:54:57.515999 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6e9cbedb-a168-495f-8746-ccf21c4e330c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cxqxn\" (UID: \"6e9cbedb-a168-495f-8746-ccf21c4e330c\") " pod="openshift-marketplace/marketplace-operator-79b997595-cxqxn" Nov 22 04:54:57 crc kubenswrapper[4818]: I1122 04:54:57.669481 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cxqxn" Nov 22 04:54:57 crc kubenswrapper[4818]: I1122 04:54:57.849077 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cxqxn"] Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.327174 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.386515 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cxqxn" event={"ID":"6e9cbedb-a168-495f-8746-ccf21c4e330c","Type":"ContainerStarted","Data":"d1a485923aee8158a0004878bbe187721352ccdda7cc06897a1c8a59eee12c8c"} Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.392545 4818 generic.go:334] "Generic (PLEG): container finished" podID="30e236a6-d269-4f06-ae6e-e3c9bb74b4bf" containerID="b00ce629adadad84a158329e4f6d5007a908430dbc8e874a6cc60a17b73dbdc2" exitCode=0 Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.392591 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg659" event={"ID":"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf","Type":"ContainerDied","Data":"b00ce629adadad84a158329e4f6d5007a908430dbc8e874a6cc60a17b73dbdc2"} Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.396405 4818 generic.go:334] "Generic (PLEG): container finished" podID="0e0614cc-57b5-4782-b462-e693c03c2117" containerID="d57a8171c3e97f6c8a2297c4ba361e5d01f7fa77f4fbcb6d72a3d2428031fa78" exitCode=0 Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.396470 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blsms" event={"ID":"0e0614cc-57b5-4782-b462-e693c03c2117","Type":"ContainerDied","Data":"d57a8171c3e97f6c8a2297c4ba361e5d01f7fa77f4fbcb6d72a3d2428031fa78"} Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.398614 4818 generic.go:334] "Generic (PLEG): container finished" podID="a48b14cb-0897-4530-8f74-bee9352bcc44" containerID="6b01728a70c0148d741315caf0a86df9846c8ca4728cbb8cbfd91db0f8c6220d" exitCode=0 Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.398656 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xg9d" event={"ID":"a48b14cb-0897-4530-8f74-bee9352bcc44","Type":"ContainerDied","Data":"6b01728a70c0148d741315caf0a86df9846c8ca4728cbb8cbfd91db0f8c6220d"} Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.401531 4818 generic.go:334] "Generic (PLEG): container finished" podID="e9310d3a-0f3d-4b29-9ab6-ea9480f40274" containerID="39298509c0016a0bef2207b2fe135731a04f371951d798aec582aeebea5e3e37" exitCode=0 Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.401592 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jswlc" event={"ID":"e9310d3a-0f3d-4b29-9ab6-ea9480f40274","Type":"ContainerDied","Data":"39298509c0016a0bef2207b2fe135731a04f371951d798aec582aeebea5e3e37"} Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.404884 4818 generic.go:334] "Generic (PLEG): container finished" podID="1d4ef859-1409-451a-a2e1-4c1c8c6d473b" containerID="1d9c4f4c749f569d569ffc735c3e3244d269e0b217dcbf2dafad509d813c3355" exitCode=0 Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.404908 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" event={"ID":"1d4ef859-1409-451a-a2e1-4c1c8c6d473b","Type":"ContainerDied","Data":"1d9c4f4c749f569d569ffc735c3e3244d269e0b217dcbf2dafad509d813c3355"} Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.404924 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" event={"ID":"1d4ef859-1409-451a-a2e1-4c1c8c6d473b","Type":"ContainerDied","Data":"609bf2f02463db58829b6c51aec6fc3185c533005fbe9a89964428fa4c396351"} Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.404941 4818 scope.go:117] "RemoveContainer" containerID="1d9c4f4c749f569d569ffc735c3e3244d269e0b217dcbf2dafad509d813c3355" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.405033 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nz99w" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.415348 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpq48\" (UniqueName: \"kubernetes.io/projected/1d4ef859-1409-451a-a2e1-4c1c8c6d473b-kube-api-access-kpq48\") pod \"1d4ef859-1409-451a-a2e1-4c1c8c6d473b\" (UID: \"1d4ef859-1409-451a-a2e1-4c1c8c6d473b\") " Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.415465 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1d4ef859-1409-451a-a2e1-4c1c8c6d473b-marketplace-trusted-ca\") pod \"1d4ef859-1409-451a-a2e1-4c1c8c6d473b\" (UID: \"1d4ef859-1409-451a-a2e1-4c1c8c6d473b\") " Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.415499 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1d4ef859-1409-451a-a2e1-4c1c8c6d473b-marketplace-operator-metrics\") pod \"1d4ef859-1409-451a-a2e1-4c1c8c6d473b\" (UID: \"1d4ef859-1409-451a-a2e1-4c1c8c6d473b\") " Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.416263 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d4ef859-1409-451a-a2e1-4c1c8c6d473b-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "1d4ef859-1409-451a-a2e1-4c1c8c6d473b" (UID: "1d4ef859-1409-451a-a2e1-4c1c8c6d473b"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.425835 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d4ef859-1409-451a-a2e1-4c1c8c6d473b-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "1d4ef859-1409-451a-a2e1-4c1c8c6d473b" (UID: "1d4ef859-1409-451a-a2e1-4c1c8c6d473b"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.428727 4818 scope.go:117] "RemoveContainer" containerID="1d9c4f4c749f569d569ffc735c3e3244d269e0b217dcbf2dafad509d813c3355" Nov 22 04:54:58 crc kubenswrapper[4818]: E1122 04:54:58.431170 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d9c4f4c749f569d569ffc735c3e3244d269e0b217dcbf2dafad509d813c3355\": container with ID starting with 1d9c4f4c749f569d569ffc735c3e3244d269e0b217dcbf2dafad509d813c3355 not found: ID does not exist" containerID="1d9c4f4c749f569d569ffc735c3e3244d269e0b217dcbf2dafad509d813c3355" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.431203 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d9c4f4c749f569d569ffc735c3e3244d269e0b217dcbf2dafad509d813c3355"} err="failed to get container status \"1d9c4f4c749f569d569ffc735c3e3244d269e0b217dcbf2dafad509d813c3355\": rpc error: code = NotFound desc = could not find container \"1d9c4f4c749f569d569ffc735c3e3244d269e0b217dcbf2dafad509d813c3355\": container with ID starting with 1d9c4f4c749f569d569ffc735c3e3244d269e0b217dcbf2dafad509d813c3355 not found: ID does not exist" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.432266 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d4ef859-1409-451a-a2e1-4c1c8c6d473b-kube-api-access-kpq48" (OuterVolumeSpecName: "kube-api-access-kpq48") pod "1d4ef859-1409-451a-a2e1-4c1c8c6d473b" (UID: "1d4ef859-1409-451a-a2e1-4c1c8c6d473b"). InnerVolumeSpecName "kube-api-access-kpq48". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.516822 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpq48\" (UniqueName: \"kubernetes.io/projected/1d4ef859-1409-451a-a2e1-4c1c8c6d473b-kube-api-access-kpq48\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.516873 4818 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1d4ef859-1409-451a-a2e1-4c1c8c6d473b-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.516886 4818 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1d4ef859-1409-451a-a2e1-4c1c8c6d473b-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.533788 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-blsms" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.567865 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pg659" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.617871 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e0614cc-57b5-4782-b462-e693c03c2117-utilities\") pod \"0e0614cc-57b5-4782-b462-e693c03c2117\" (UID: \"0e0614cc-57b5-4782-b462-e693c03c2117\") " Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.617939 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stfxm\" (UniqueName: \"kubernetes.io/projected/30e236a6-d269-4f06-ae6e-e3c9bb74b4bf-kube-api-access-stfxm\") pod \"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf\" (UID: \"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf\") " Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.618009 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e0614cc-57b5-4782-b462-e693c03c2117-catalog-content\") pod \"0e0614cc-57b5-4782-b462-e693c03c2117\" (UID: \"0e0614cc-57b5-4782-b462-e693c03c2117\") " Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.618031 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bclj5\" (UniqueName: \"kubernetes.io/projected/0e0614cc-57b5-4782-b462-e693c03c2117-kube-api-access-bclj5\") pod \"0e0614cc-57b5-4782-b462-e693c03c2117\" (UID: \"0e0614cc-57b5-4782-b462-e693c03c2117\") " Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.618092 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30e236a6-d269-4f06-ae6e-e3c9bb74b4bf-utilities\") pod \"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf\" (UID: \"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf\") " Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.618113 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30e236a6-d269-4f06-ae6e-e3c9bb74b4bf-catalog-content\") pod \"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf\" (UID: \"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf\") " Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.618743 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e0614cc-57b5-4782-b462-e693c03c2117-utilities" (OuterVolumeSpecName: "utilities") pod "0e0614cc-57b5-4782-b462-e693c03c2117" (UID: "0e0614cc-57b5-4782-b462-e693c03c2117"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.619207 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e0614cc-57b5-4782-b462-e693c03c2117-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.619235 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30e236a6-d269-4f06-ae6e-e3c9bb74b4bf-utilities" (OuterVolumeSpecName: "utilities") pod "30e236a6-d269-4f06-ae6e-e3c9bb74b4bf" (UID: "30e236a6-d269-4f06-ae6e-e3c9bb74b4bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.622937 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30e236a6-d269-4f06-ae6e-e3c9bb74b4bf-kube-api-access-stfxm" (OuterVolumeSpecName: "kube-api-access-stfxm") pod "30e236a6-d269-4f06-ae6e-e3c9bb74b4bf" (UID: "30e236a6-d269-4f06-ae6e-e3c9bb74b4bf"). InnerVolumeSpecName "kube-api-access-stfxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.623103 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e0614cc-57b5-4782-b462-e693c03c2117-kube-api-access-bclj5" (OuterVolumeSpecName: "kube-api-access-bclj5") pod "0e0614cc-57b5-4782-b462-e693c03c2117" (UID: "0e0614cc-57b5-4782-b462-e693c03c2117"). InnerVolumeSpecName "kube-api-access-bclj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.660168 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7xg9d" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.671634 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jswlc" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.679064 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30e236a6-d269-4f06-ae6e-e3c9bb74b4bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30e236a6-d269-4f06-ae6e-e3c9bb74b4bf" (UID: "30e236a6-d269-4f06-ae6e-e3c9bb74b4bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.699194 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e0614cc-57b5-4782-b462-e693c03c2117-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0e0614cc-57b5-4782-b462-e693c03c2117" (UID: "0e0614cc-57b5-4782-b462-e693c03c2117"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.720192 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a48b14cb-0897-4530-8f74-bee9352bcc44-catalog-content\") pod \"a48b14cb-0897-4530-8f74-bee9352bcc44\" (UID: \"a48b14cb-0897-4530-8f74-bee9352bcc44\") " Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.720291 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcjvh\" (UniqueName: \"kubernetes.io/projected/a48b14cb-0897-4530-8f74-bee9352bcc44-kube-api-access-bcjvh\") pod \"a48b14cb-0897-4530-8f74-bee9352bcc44\" (UID: \"a48b14cb-0897-4530-8f74-bee9352bcc44\") " Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.720330 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9310d3a-0f3d-4b29-9ab6-ea9480f40274-catalog-content\") pod \"e9310d3a-0f3d-4b29-9ab6-ea9480f40274\" (UID: \"e9310d3a-0f3d-4b29-9ab6-ea9480f40274\") " Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.720376 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shjjf\" (UniqueName: \"kubernetes.io/projected/e9310d3a-0f3d-4b29-9ab6-ea9480f40274-kube-api-access-shjjf\") pod \"e9310d3a-0f3d-4b29-9ab6-ea9480f40274\" (UID: \"e9310d3a-0f3d-4b29-9ab6-ea9480f40274\") " Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.720398 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a48b14cb-0897-4530-8f74-bee9352bcc44-utilities\") pod \"a48b14cb-0897-4530-8f74-bee9352bcc44\" (UID: \"a48b14cb-0897-4530-8f74-bee9352bcc44\") " Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.720437 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9310d3a-0f3d-4b29-9ab6-ea9480f40274-utilities\") pod \"e9310d3a-0f3d-4b29-9ab6-ea9480f40274\" (UID: \"e9310d3a-0f3d-4b29-9ab6-ea9480f40274\") " Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.720653 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e0614cc-57b5-4782-b462-e693c03c2117-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.720669 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bclj5\" (UniqueName: \"kubernetes.io/projected/0e0614cc-57b5-4782-b462-e693c03c2117-kube-api-access-bclj5\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.720678 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30e236a6-d269-4f06-ae6e-e3c9bb74b4bf-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.720686 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30e236a6-d269-4f06-ae6e-e3c9bb74b4bf-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.720695 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stfxm\" (UniqueName: \"kubernetes.io/projected/30e236a6-d269-4f06-ae6e-e3c9bb74b4bf-kube-api-access-stfxm\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.721408 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9310d3a-0f3d-4b29-9ab6-ea9480f40274-utilities" (OuterVolumeSpecName: "utilities") pod "e9310d3a-0f3d-4b29-9ab6-ea9480f40274" (UID: "e9310d3a-0f3d-4b29-9ab6-ea9480f40274"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.730480 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a48b14cb-0897-4530-8f74-bee9352bcc44-utilities" (OuterVolumeSpecName: "utilities") pod "a48b14cb-0897-4530-8f74-bee9352bcc44" (UID: "a48b14cb-0897-4530-8f74-bee9352bcc44"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.731409 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a48b14cb-0897-4530-8f74-bee9352bcc44-kube-api-access-bcjvh" (OuterVolumeSpecName: "kube-api-access-bcjvh") pod "a48b14cb-0897-4530-8f74-bee9352bcc44" (UID: "a48b14cb-0897-4530-8f74-bee9352bcc44"). InnerVolumeSpecName "kube-api-access-bcjvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.731473 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nz99w"] Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.731468 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9310d3a-0f3d-4b29-9ab6-ea9480f40274-kube-api-access-shjjf" (OuterVolumeSpecName: "kube-api-access-shjjf") pod "e9310d3a-0f3d-4b29-9ab6-ea9480f40274" (UID: "e9310d3a-0f3d-4b29-9ab6-ea9480f40274"). InnerVolumeSpecName "kube-api-access-shjjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.733815 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nz99w"] Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.750994 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9310d3a-0f3d-4b29-9ab6-ea9480f40274-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e9310d3a-0f3d-4b29-9ab6-ea9480f40274" (UID: "e9310d3a-0f3d-4b29-9ab6-ea9480f40274"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.817304 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a48b14cb-0897-4530-8f74-bee9352bcc44-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a48b14cb-0897-4530-8f74-bee9352bcc44" (UID: "a48b14cb-0897-4530-8f74-bee9352bcc44"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.822347 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a48b14cb-0897-4530-8f74-bee9352bcc44-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.822385 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcjvh\" (UniqueName: \"kubernetes.io/projected/a48b14cb-0897-4530-8f74-bee9352bcc44-kube-api-access-bcjvh\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.822396 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9310d3a-0f3d-4b29-9ab6-ea9480f40274-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.822405 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shjjf\" (UniqueName: \"kubernetes.io/projected/e9310d3a-0f3d-4b29-9ab6-ea9480f40274-kube-api-access-shjjf\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.822415 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a48b14cb-0897-4530-8f74-bee9352bcc44-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:58 crc kubenswrapper[4818]: I1122 04:54:58.822422 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9310d3a-0f3d-4b29-9ab6-ea9480f40274-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.412897 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blsms" event={"ID":"0e0614cc-57b5-4782-b462-e693c03c2117","Type":"ContainerDied","Data":"1568c4aee9967dab35f69933192e224daed6ade8e9de040b5fa5a275d1664f09"} Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.412972 4818 scope.go:117] "RemoveContainer" containerID="d57a8171c3e97f6c8a2297c4ba361e5d01f7fa77f4fbcb6d72a3d2428031fa78" Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.412963 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-blsms" Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.416432 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xg9d" event={"ID":"a48b14cb-0897-4530-8f74-bee9352bcc44","Type":"ContainerDied","Data":"60e27568ebe4434e6e4270b1d6154ccac213e437223ca96fd461627d16d5b119"} Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.416611 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7xg9d" Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.421171 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jswlc" Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.421199 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jswlc" event={"ID":"e9310d3a-0f3d-4b29-9ab6-ea9480f40274","Type":"ContainerDied","Data":"bb765bd3f43771442f8485d583573f762bfa41288f71f3732719b7087f819eae"} Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.423991 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cxqxn" event={"ID":"6e9cbedb-a168-495f-8746-ccf21c4e330c","Type":"ContainerStarted","Data":"7a14de2575999a31e4afde0db764607016e2bae46298124aad4a2d57ccb0b955"} Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.425243 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-cxqxn" Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.427745 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg659" event={"ID":"30e236a6-d269-4f06-ae6e-e3c9bb74b4bf","Type":"ContainerDied","Data":"2648717bb6765491289df8b05a194e3c501ac1fcf258a4c38e8d3fc7f024f27a"} Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.427818 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pg659" Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.435093 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-cxqxn" Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.448113 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-cxqxn" podStartSLOduration=3.448019599 podStartE2EDuration="3.448019599s" podCreationTimestamp="2025-11-22 04:54:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:54:59.444416169 +0000 UTC m=+452.018832696" watchObservedRunningTime="2025-11-22 04:54:59.448019599 +0000 UTC m=+452.022436156" Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.453244 4818 scope.go:117] "RemoveContainer" containerID="e8da1f9fbc283ea63d1501ce1f463e25a8f883ba1be36bf9cc7cf0f7f095aee2" Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.467239 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jswlc"] Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.472834 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jswlc"] Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.490342 4818 scope.go:117] "RemoveContainer" containerID="bc26224b836a30445b253579c27f8308345be71519b52f86ecd93d4240ea27ff" Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.504886 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-blsms"] Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.507502 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-blsms"] Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.511619 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7xg9d"] Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.516555 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7xg9d"] Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.525554 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pg659"] Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.526158 4818 scope.go:117] "RemoveContainer" containerID="6b01728a70c0148d741315caf0a86df9846c8ca4728cbb8cbfd91db0f8c6220d" Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.527466 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pg659"] Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.544539 4818 scope.go:117] "RemoveContainer" containerID="4b0cbc23059764fad28d1e357825d9bb4c667b16a2fdae1c3c2828111669d1c5" Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.566413 4818 scope.go:117] "RemoveContainer" containerID="6750ee30fbe98f95438522cf05d598dff7b0e291eb14575aa75748e4e558cbfe" Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.581117 4818 scope.go:117] "RemoveContainer" containerID="39298509c0016a0bef2207b2fe135731a04f371951d798aec582aeebea5e3e37" Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.599340 4818 scope.go:117] "RemoveContainer" containerID="523bb3e9efd24499c8c60fdbbdd08f33ed2d49d3e16128898f98afd11cb1777f" Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.620477 4818 scope.go:117] "RemoveContainer" containerID="d2c01cd7b5ac31c71cd8423201f20a33b1f945ecba8213ad5506cf2b3290c40c" Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.631467 4818 scope.go:117] "RemoveContainer" containerID="b00ce629adadad84a158329e4f6d5007a908430dbc8e874a6cc60a17b73dbdc2" Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.641396 4818 scope.go:117] "RemoveContainer" containerID="d7d124f9a3f300c251e85323f4282b72f96feeda714da160d9fd4a426c3f5e0e" Nov 22 04:54:59 crc kubenswrapper[4818]: I1122 04:54:59.656051 4818 scope.go:117] "RemoveContainer" containerID="4e09d0edf82e0b5cf5fdac97850852843dfac2b775e16e47c9025a816ecb4d62" Nov 22 04:55:00 crc kubenswrapper[4818]: I1122 04:55:00.299860 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e0614cc-57b5-4782-b462-e693c03c2117" path="/var/lib/kubelet/pods/0e0614cc-57b5-4782-b462-e693c03c2117/volumes" Nov 22 04:55:00 crc kubenswrapper[4818]: I1122 04:55:00.300485 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d4ef859-1409-451a-a2e1-4c1c8c6d473b" path="/var/lib/kubelet/pods/1d4ef859-1409-451a-a2e1-4c1c8c6d473b/volumes" Nov 22 04:55:00 crc kubenswrapper[4818]: I1122 04:55:00.300932 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30e236a6-d269-4f06-ae6e-e3c9bb74b4bf" path="/var/lib/kubelet/pods/30e236a6-d269-4f06-ae6e-e3c9bb74b4bf/volumes" Nov 22 04:55:00 crc kubenswrapper[4818]: I1122 04:55:00.301478 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a48b14cb-0897-4530-8f74-bee9352bcc44" path="/var/lib/kubelet/pods/a48b14cb-0897-4530-8f74-bee9352bcc44/volumes" Nov 22 04:55:00 crc kubenswrapper[4818]: I1122 04:55:00.302716 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9310d3a-0f3d-4b29-9ab6-ea9480f40274" path="/var/lib/kubelet/pods/e9310d3a-0f3d-4b29-9ab6-ea9480f40274/volumes" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.018361 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vcnvz"] Nov 22 04:55:01 crc kubenswrapper[4818]: E1122 04:55:01.018596 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30e236a6-d269-4f06-ae6e-e3c9bb74b4bf" containerName="extract-content" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.018607 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="30e236a6-d269-4f06-ae6e-e3c9bb74b4bf" containerName="extract-content" Nov 22 04:55:01 crc kubenswrapper[4818]: E1122 04:55:01.018620 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e0614cc-57b5-4782-b462-e693c03c2117" containerName="extract-utilities" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.018625 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e0614cc-57b5-4782-b462-e693c03c2117" containerName="extract-utilities" Nov 22 04:55:01 crc kubenswrapper[4818]: E1122 04:55:01.018634 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9310d3a-0f3d-4b29-9ab6-ea9480f40274" containerName="extract-utilities" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.018639 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9310d3a-0f3d-4b29-9ab6-ea9480f40274" containerName="extract-utilities" Nov 22 04:55:01 crc kubenswrapper[4818]: E1122 04:55:01.018647 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e0614cc-57b5-4782-b462-e693c03c2117" containerName="registry-server" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.018653 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e0614cc-57b5-4782-b462-e693c03c2117" containerName="registry-server" Nov 22 04:55:01 crc kubenswrapper[4818]: E1122 04:55:01.018660 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a48b14cb-0897-4530-8f74-bee9352bcc44" containerName="extract-utilities" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.018666 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a48b14cb-0897-4530-8f74-bee9352bcc44" containerName="extract-utilities" Nov 22 04:55:01 crc kubenswrapper[4818]: E1122 04:55:01.018672 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d4ef859-1409-451a-a2e1-4c1c8c6d473b" containerName="marketplace-operator" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.018678 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d4ef859-1409-451a-a2e1-4c1c8c6d473b" containerName="marketplace-operator" Nov 22 04:55:01 crc kubenswrapper[4818]: E1122 04:55:01.018687 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a48b14cb-0897-4530-8f74-bee9352bcc44" containerName="registry-server" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.018694 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a48b14cb-0897-4530-8f74-bee9352bcc44" containerName="registry-server" Nov 22 04:55:01 crc kubenswrapper[4818]: E1122 04:55:01.018700 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9310d3a-0f3d-4b29-9ab6-ea9480f40274" containerName="registry-server" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.018705 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9310d3a-0f3d-4b29-9ab6-ea9480f40274" containerName="registry-server" Nov 22 04:55:01 crc kubenswrapper[4818]: E1122 04:55:01.018711 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9310d3a-0f3d-4b29-9ab6-ea9480f40274" containerName="extract-content" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.018717 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9310d3a-0f3d-4b29-9ab6-ea9480f40274" containerName="extract-content" Nov 22 04:55:01 crc kubenswrapper[4818]: E1122 04:55:01.018725 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30e236a6-d269-4f06-ae6e-e3c9bb74b4bf" containerName="extract-utilities" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.018731 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="30e236a6-d269-4f06-ae6e-e3c9bb74b4bf" containerName="extract-utilities" Nov 22 04:55:01 crc kubenswrapper[4818]: E1122 04:55:01.018738 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a48b14cb-0897-4530-8f74-bee9352bcc44" containerName="extract-content" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.018743 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a48b14cb-0897-4530-8f74-bee9352bcc44" containerName="extract-content" Nov 22 04:55:01 crc kubenswrapper[4818]: E1122 04:55:01.018750 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30e236a6-d269-4f06-ae6e-e3c9bb74b4bf" containerName="registry-server" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.018755 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="30e236a6-d269-4f06-ae6e-e3c9bb74b4bf" containerName="registry-server" Nov 22 04:55:01 crc kubenswrapper[4818]: E1122 04:55:01.018765 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e0614cc-57b5-4782-b462-e693c03c2117" containerName="extract-content" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.018771 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e0614cc-57b5-4782-b462-e693c03c2117" containerName="extract-content" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.018854 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9310d3a-0f3d-4b29-9ab6-ea9480f40274" containerName="registry-server" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.018864 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="30e236a6-d269-4f06-ae6e-e3c9bb74b4bf" containerName="registry-server" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.018871 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e0614cc-57b5-4782-b462-e693c03c2117" containerName="registry-server" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.018878 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a48b14cb-0897-4530-8f74-bee9352bcc44" containerName="registry-server" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.018886 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d4ef859-1409-451a-a2e1-4c1c8c6d473b" containerName="marketplace-operator" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.019544 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vcnvz" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.022124 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.031899 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vcnvz"] Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.044683 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/603888a5-b8ab-4378-ab59-d044221e403c-catalog-content\") pod \"certified-operators-vcnvz\" (UID: \"603888a5-b8ab-4378-ab59-d044221e403c\") " pod="openshift-marketplace/certified-operators-vcnvz" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.044755 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/603888a5-b8ab-4378-ab59-d044221e403c-utilities\") pod \"certified-operators-vcnvz\" (UID: \"603888a5-b8ab-4378-ab59-d044221e403c\") " pod="openshift-marketplace/certified-operators-vcnvz" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.044954 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc7k8\" (UniqueName: \"kubernetes.io/projected/603888a5-b8ab-4378-ab59-d044221e403c-kube-api-access-pc7k8\") pod \"certified-operators-vcnvz\" (UID: \"603888a5-b8ab-4378-ab59-d044221e403c\") " pod="openshift-marketplace/certified-operators-vcnvz" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.145783 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/603888a5-b8ab-4378-ab59-d044221e403c-catalog-content\") pod \"certified-operators-vcnvz\" (UID: \"603888a5-b8ab-4378-ab59-d044221e403c\") " pod="openshift-marketplace/certified-operators-vcnvz" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.145849 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/603888a5-b8ab-4378-ab59-d044221e403c-utilities\") pod \"certified-operators-vcnvz\" (UID: \"603888a5-b8ab-4378-ab59-d044221e403c\") " pod="openshift-marketplace/certified-operators-vcnvz" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.145905 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc7k8\" (UniqueName: \"kubernetes.io/projected/603888a5-b8ab-4378-ab59-d044221e403c-kube-api-access-pc7k8\") pod \"certified-operators-vcnvz\" (UID: \"603888a5-b8ab-4378-ab59-d044221e403c\") " pod="openshift-marketplace/certified-operators-vcnvz" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.146899 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/603888a5-b8ab-4378-ab59-d044221e403c-utilities\") pod \"certified-operators-vcnvz\" (UID: \"603888a5-b8ab-4378-ab59-d044221e403c\") " pod="openshift-marketplace/certified-operators-vcnvz" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.146917 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/603888a5-b8ab-4378-ab59-d044221e403c-catalog-content\") pod \"certified-operators-vcnvz\" (UID: \"603888a5-b8ab-4378-ab59-d044221e403c\") " pod="openshift-marketplace/certified-operators-vcnvz" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.165046 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc7k8\" (UniqueName: \"kubernetes.io/projected/603888a5-b8ab-4378-ab59-d044221e403c-kube-api-access-pc7k8\") pod \"certified-operators-vcnvz\" (UID: \"603888a5-b8ab-4378-ab59-d044221e403c\") " pod="openshift-marketplace/certified-operators-vcnvz" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.219001 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k4j4l"] Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.219998 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k4j4l" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.222548 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.235702 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k4j4l"] Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.248581 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bggh\" (UniqueName: \"kubernetes.io/projected/614051f2-a9cf-4084-9eb0-d5333b5b1d0e-kube-api-access-8bggh\") pod \"community-operators-k4j4l\" (UID: \"614051f2-a9cf-4084-9eb0-d5333b5b1d0e\") " pod="openshift-marketplace/community-operators-k4j4l" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.248691 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/614051f2-a9cf-4084-9eb0-d5333b5b1d0e-catalog-content\") pod \"community-operators-k4j4l\" (UID: \"614051f2-a9cf-4084-9eb0-d5333b5b1d0e\") " pod="openshift-marketplace/community-operators-k4j4l" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.248719 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/614051f2-a9cf-4084-9eb0-d5333b5b1d0e-utilities\") pod \"community-operators-k4j4l\" (UID: \"614051f2-a9cf-4084-9eb0-d5333b5b1d0e\") " pod="openshift-marketplace/community-operators-k4j4l" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.349814 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/614051f2-a9cf-4084-9eb0-d5333b5b1d0e-utilities\") pod \"community-operators-k4j4l\" (UID: \"614051f2-a9cf-4084-9eb0-d5333b5b1d0e\") " pod="openshift-marketplace/community-operators-k4j4l" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.350158 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bggh\" (UniqueName: \"kubernetes.io/projected/614051f2-a9cf-4084-9eb0-d5333b5b1d0e-kube-api-access-8bggh\") pod \"community-operators-k4j4l\" (UID: \"614051f2-a9cf-4084-9eb0-d5333b5b1d0e\") " pod="openshift-marketplace/community-operators-k4j4l" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.350245 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/614051f2-a9cf-4084-9eb0-d5333b5b1d0e-catalog-content\") pod \"community-operators-k4j4l\" (UID: \"614051f2-a9cf-4084-9eb0-d5333b5b1d0e\") " pod="openshift-marketplace/community-operators-k4j4l" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.350724 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/614051f2-a9cf-4084-9eb0-d5333b5b1d0e-catalog-content\") pod \"community-operators-k4j4l\" (UID: \"614051f2-a9cf-4084-9eb0-d5333b5b1d0e\") " pod="openshift-marketplace/community-operators-k4j4l" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.351636 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/614051f2-a9cf-4084-9eb0-d5333b5b1d0e-utilities\") pod \"community-operators-k4j4l\" (UID: \"614051f2-a9cf-4084-9eb0-d5333b5b1d0e\") " pod="openshift-marketplace/community-operators-k4j4l" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.369610 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bggh\" (UniqueName: \"kubernetes.io/projected/614051f2-a9cf-4084-9eb0-d5333b5b1d0e-kube-api-access-8bggh\") pod \"community-operators-k4j4l\" (UID: \"614051f2-a9cf-4084-9eb0-d5333b5b1d0e\") " pod="openshift-marketplace/community-operators-k4j4l" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.401131 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vcnvz" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.534822 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k4j4l" Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.610048 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vcnvz"] Nov 22 04:55:01 crc kubenswrapper[4818]: I1122 04:55:01.767593 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k4j4l"] Nov 22 04:55:01 crc kubenswrapper[4818]: W1122 04:55:01.774090 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod614051f2_a9cf_4084_9eb0_d5333b5b1d0e.slice/crio-a033544b5969fa584315e0a96787c02ac6d21d90a2eb00e59381f6c76e42cfe6 WatchSource:0}: Error finding container a033544b5969fa584315e0a96787c02ac6d21d90a2eb00e59381f6c76e42cfe6: Status 404 returned error can't find the container with id a033544b5969fa584315e0a96787c02ac6d21d90a2eb00e59381f6c76e42cfe6 Nov 22 04:55:02 crc kubenswrapper[4818]: I1122 04:55:02.446778 4818 generic.go:334] "Generic (PLEG): container finished" podID="603888a5-b8ab-4378-ab59-d044221e403c" containerID="1cbcafb2905f946a8cff30eb8ea6fea0c49bae176a48c9bdf2e8732cbea87a3e" exitCode=0 Nov 22 04:55:02 crc kubenswrapper[4818]: I1122 04:55:02.446838 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vcnvz" event={"ID":"603888a5-b8ab-4378-ab59-d044221e403c","Type":"ContainerDied","Data":"1cbcafb2905f946a8cff30eb8ea6fea0c49bae176a48c9bdf2e8732cbea87a3e"} Nov 22 04:55:02 crc kubenswrapper[4818]: I1122 04:55:02.447051 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vcnvz" event={"ID":"603888a5-b8ab-4378-ab59-d044221e403c","Type":"ContainerStarted","Data":"57c718accaffb09b3887635aa7ecbec356ef708da106ee2421830d8f052530cf"} Nov 22 04:55:02 crc kubenswrapper[4818]: I1122 04:55:02.451033 4818 generic.go:334] "Generic (PLEG): container finished" podID="614051f2-a9cf-4084-9eb0-d5333b5b1d0e" containerID="363e7563ae61d8f0492a8522f0937ba006e388d8d0b2cdf2d5c744363c7a55aa" exitCode=0 Nov 22 04:55:02 crc kubenswrapper[4818]: I1122 04:55:02.451068 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4j4l" event={"ID":"614051f2-a9cf-4084-9eb0-d5333b5b1d0e","Type":"ContainerDied","Data":"363e7563ae61d8f0492a8522f0937ba006e388d8d0b2cdf2d5c744363c7a55aa"} Nov 22 04:55:02 crc kubenswrapper[4818]: I1122 04:55:02.451092 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4j4l" event={"ID":"614051f2-a9cf-4084-9eb0-d5333b5b1d0e","Type":"ContainerStarted","Data":"a033544b5969fa584315e0a96787c02ac6d21d90a2eb00e59381f6c76e42cfe6"} Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.419071 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rcbjp"] Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.420873 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rcbjp" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.422802 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.432798 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rcbjp"] Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.457066 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4j4l" event={"ID":"614051f2-a9cf-4084-9eb0-d5333b5b1d0e","Type":"ContainerStarted","Data":"2cee89497b77ac47a9bf107b7d4c048076db2e5171713d3612ed4d2831d7ed0c"} Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.483508 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d83d4bfc-8842-4d00-96d9-779a89080441-catalog-content\") pod \"redhat-marketplace-rcbjp\" (UID: \"d83d4bfc-8842-4d00-96d9-779a89080441\") " pod="openshift-marketplace/redhat-marketplace-rcbjp" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.483625 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnjx9\" (UniqueName: \"kubernetes.io/projected/d83d4bfc-8842-4d00-96d9-779a89080441-kube-api-access-fnjx9\") pod \"redhat-marketplace-rcbjp\" (UID: \"d83d4bfc-8842-4d00-96d9-779a89080441\") " pod="openshift-marketplace/redhat-marketplace-rcbjp" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.483648 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d83d4bfc-8842-4d00-96d9-779a89080441-utilities\") pod \"redhat-marketplace-rcbjp\" (UID: \"d83d4bfc-8842-4d00-96d9-779a89080441\") " pod="openshift-marketplace/redhat-marketplace-rcbjp" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.587818 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnjx9\" (UniqueName: \"kubernetes.io/projected/d83d4bfc-8842-4d00-96d9-779a89080441-kube-api-access-fnjx9\") pod \"redhat-marketplace-rcbjp\" (UID: \"d83d4bfc-8842-4d00-96d9-779a89080441\") " pod="openshift-marketplace/redhat-marketplace-rcbjp" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.587871 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d83d4bfc-8842-4d00-96d9-779a89080441-utilities\") pod \"redhat-marketplace-rcbjp\" (UID: \"d83d4bfc-8842-4d00-96d9-779a89080441\") " pod="openshift-marketplace/redhat-marketplace-rcbjp" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.587906 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d83d4bfc-8842-4d00-96d9-779a89080441-catalog-content\") pod \"redhat-marketplace-rcbjp\" (UID: \"d83d4bfc-8842-4d00-96d9-779a89080441\") " pod="openshift-marketplace/redhat-marketplace-rcbjp" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.588444 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d83d4bfc-8842-4d00-96d9-779a89080441-catalog-content\") pod \"redhat-marketplace-rcbjp\" (UID: \"d83d4bfc-8842-4d00-96d9-779a89080441\") " pod="openshift-marketplace/redhat-marketplace-rcbjp" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.588467 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d83d4bfc-8842-4d00-96d9-779a89080441-utilities\") pod \"redhat-marketplace-rcbjp\" (UID: \"d83d4bfc-8842-4d00-96d9-779a89080441\") " pod="openshift-marketplace/redhat-marketplace-rcbjp" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.614102 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnjx9\" (UniqueName: \"kubernetes.io/projected/d83d4bfc-8842-4d00-96d9-779a89080441-kube-api-access-fnjx9\") pod \"redhat-marketplace-rcbjp\" (UID: \"d83d4bfc-8842-4d00-96d9-779a89080441\") " pod="openshift-marketplace/redhat-marketplace-rcbjp" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.626348 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h8lq9"] Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.629109 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h8lq9" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.630600 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h8lq9"] Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.630988 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.689030 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/759a3ee8-f990-4960-bc61-9893ca8cbc6f-catalog-content\") pod \"redhat-operators-h8lq9\" (UID: \"759a3ee8-f990-4960-bc61-9893ca8cbc6f\") " pod="openshift-marketplace/redhat-operators-h8lq9" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.689117 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7zr9\" (UniqueName: \"kubernetes.io/projected/759a3ee8-f990-4960-bc61-9893ca8cbc6f-kube-api-access-z7zr9\") pod \"redhat-operators-h8lq9\" (UID: \"759a3ee8-f990-4960-bc61-9893ca8cbc6f\") " pod="openshift-marketplace/redhat-operators-h8lq9" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.689142 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/759a3ee8-f990-4960-bc61-9893ca8cbc6f-utilities\") pod \"redhat-operators-h8lq9\" (UID: \"759a3ee8-f990-4960-bc61-9893ca8cbc6f\") " pod="openshift-marketplace/redhat-operators-h8lq9" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.747662 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rcbjp" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.790309 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7zr9\" (UniqueName: \"kubernetes.io/projected/759a3ee8-f990-4960-bc61-9893ca8cbc6f-kube-api-access-z7zr9\") pod \"redhat-operators-h8lq9\" (UID: \"759a3ee8-f990-4960-bc61-9893ca8cbc6f\") " pod="openshift-marketplace/redhat-operators-h8lq9" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.790366 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/759a3ee8-f990-4960-bc61-9893ca8cbc6f-utilities\") pod \"redhat-operators-h8lq9\" (UID: \"759a3ee8-f990-4960-bc61-9893ca8cbc6f\") " pod="openshift-marketplace/redhat-operators-h8lq9" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.790410 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/759a3ee8-f990-4960-bc61-9893ca8cbc6f-catalog-content\") pod \"redhat-operators-h8lq9\" (UID: \"759a3ee8-f990-4960-bc61-9893ca8cbc6f\") " pod="openshift-marketplace/redhat-operators-h8lq9" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.790779 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/759a3ee8-f990-4960-bc61-9893ca8cbc6f-utilities\") pod \"redhat-operators-h8lq9\" (UID: \"759a3ee8-f990-4960-bc61-9893ca8cbc6f\") " pod="openshift-marketplace/redhat-operators-h8lq9" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.790837 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/759a3ee8-f990-4960-bc61-9893ca8cbc6f-catalog-content\") pod \"redhat-operators-h8lq9\" (UID: \"759a3ee8-f990-4960-bc61-9893ca8cbc6f\") " pod="openshift-marketplace/redhat-operators-h8lq9" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.810733 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7zr9\" (UniqueName: \"kubernetes.io/projected/759a3ee8-f990-4960-bc61-9893ca8cbc6f-kube-api-access-z7zr9\") pod \"redhat-operators-h8lq9\" (UID: \"759a3ee8-f990-4960-bc61-9893ca8cbc6f\") " pod="openshift-marketplace/redhat-operators-h8lq9" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.932087 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rcbjp"] Nov 22 04:55:03 crc kubenswrapper[4818]: W1122 04:55:03.947045 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd83d4bfc_8842_4d00_96d9_779a89080441.slice/crio-a18621c31c9aa6052d625855131eee9f4b182cb5151c1800bcdc83c63682150f WatchSource:0}: Error finding container a18621c31c9aa6052d625855131eee9f4b182cb5151c1800bcdc83c63682150f: Status 404 returned error can't find the container with id a18621c31c9aa6052d625855131eee9f4b182cb5151c1800bcdc83c63682150f Nov 22 04:55:03 crc kubenswrapper[4818]: E1122 04:55:03.954217 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod603888a5_b8ab_4378_ab59_d044221e403c.slice/crio-638b7ce27977086e227aac8cbfea309b16c22ba935354eb259bd615bfae86f5a.scope\": RecentStats: unable to find data in memory cache]" Nov 22 04:55:03 crc kubenswrapper[4818]: I1122 04:55:03.959760 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h8lq9" Nov 22 04:55:04 crc kubenswrapper[4818]: I1122 04:55:04.139724 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h8lq9"] Nov 22 04:55:04 crc kubenswrapper[4818]: I1122 04:55:04.464431 4818 generic.go:334] "Generic (PLEG): container finished" podID="603888a5-b8ab-4378-ab59-d044221e403c" containerID="638b7ce27977086e227aac8cbfea309b16c22ba935354eb259bd615bfae86f5a" exitCode=0 Nov 22 04:55:04 crc kubenswrapper[4818]: I1122 04:55:04.464484 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vcnvz" event={"ID":"603888a5-b8ab-4378-ab59-d044221e403c","Type":"ContainerDied","Data":"638b7ce27977086e227aac8cbfea309b16c22ba935354eb259bd615bfae86f5a"} Nov 22 04:55:04 crc kubenswrapper[4818]: I1122 04:55:04.466750 4818 generic.go:334] "Generic (PLEG): container finished" podID="759a3ee8-f990-4960-bc61-9893ca8cbc6f" containerID="3c20c1959c9631fd5af308e1b23eb72c0ed16529783c5e8935176e4456a9ea67" exitCode=0 Nov 22 04:55:04 crc kubenswrapper[4818]: I1122 04:55:04.466847 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8lq9" event={"ID":"759a3ee8-f990-4960-bc61-9893ca8cbc6f","Type":"ContainerDied","Data":"3c20c1959c9631fd5af308e1b23eb72c0ed16529783c5e8935176e4456a9ea67"} Nov 22 04:55:04 crc kubenswrapper[4818]: I1122 04:55:04.466877 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8lq9" event={"ID":"759a3ee8-f990-4960-bc61-9893ca8cbc6f","Type":"ContainerStarted","Data":"93544e8634fe97878d772d4b092b3c32c8efc6d781655c58d1e768cf5a361195"} Nov 22 04:55:04 crc kubenswrapper[4818]: I1122 04:55:04.468895 4818 generic.go:334] "Generic (PLEG): container finished" podID="614051f2-a9cf-4084-9eb0-d5333b5b1d0e" containerID="2cee89497b77ac47a9bf107b7d4c048076db2e5171713d3612ed4d2831d7ed0c" exitCode=0 Nov 22 04:55:04 crc kubenswrapper[4818]: I1122 04:55:04.468965 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4j4l" event={"ID":"614051f2-a9cf-4084-9eb0-d5333b5b1d0e","Type":"ContainerDied","Data":"2cee89497b77ac47a9bf107b7d4c048076db2e5171713d3612ed4d2831d7ed0c"} Nov 22 04:55:04 crc kubenswrapper[4818]: I1122 04:55:04.470523 4818 generic.go:334] "Generic (PLEG): container finished" podID="d83d4bfc-8842-4d00-96d9-779a89080441" containerID="673707dc837f59a6ba7c79647116f88c1a222706957d4a0ace55d0b0b839a807" exitCode=0 Nov 22 04:55:04 crc kubenswrapper[4818]: I1122 04:55:04.470588 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rcbjp" event={"ID":"d83d4bfc-8842-4d00-96d9-779a89080441","Type":"ContainerDied","Data":"673707dc837f59a6ba7c79647116f88c1a222706957d4a0ace55d0b0b839a807"} Nov 22 04:55:04 crc kubenswrapper[4818]: I1122 04:55:04.470685 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rcbjp" event={"ID":"d83d4bfc-8842-4d00-96d9-779a89080441","Type":"ContainerStarted","Data":"a18621c31c9aa6052d625855131eee9f4b182cb5151c1800bcdc83c63682150f"} Nov 22 04:55:06 crc kubenswrapper[4818]: I1122 04:55:06.485513 4818 generic.go:334] "Generic (PLEG): container finished" podID="759a3ee8-f990-4960-bc61-9893ca8cbc6f" containerID="4e601462824ee078eed9cb485075f4ff0a45f57247b349b22540bc25498ba588" exitCode=0 Nov 22 04:55:06 crc kubenswrapper[4818]: I1122 04:55:06.485709 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8lq9" event={"ID":"759a3ee8-f990-4960-bc61-9893ca8cbc6f","Type":"ContainerDied","Data":"4e601462824ee078eed9cb485075f4ff0a45f57247b349b22540bc25498ba588"} Nov 22 04:55:06 crc kubenswrapper[4818]: I1122 04:55:06.487624 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 04:55:06 crc kubenswrapper[4818]: I1122 04:55:06.488546 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4j4l" event={"ID":"614051f2-a9cf-4084-9eb0-d5333b5b1d0e","Type":"ContainerStarted","Data":"a4886460724ba5684813c21db17c3989575dea69dff6dcd1076005977210e193"} Nov 22 04:55:06 crc kubenswrapper[4818]: I1122 04:55:06.490863 4818 generic.go:334] "Generic (PLEG): container finished" podID="d83d4bfc-8842-4d00-96d9-779a89080441" containerID="013858c6210de2aad706d02bf3974b0bbaa6f15fb05920a786b025cb555461b6" exitCode=0 Nov 22 04:55:06 crc kubenswrapper[4818]: I1122 04:55:06.490895 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rcbjp" event={"ID":"d83d4bfc-8842-4d00-96d9-779a89080441","Type":"ContainerDied","Data":"013858c6210de2aad706d02bf3974b0bbaa6f15fb05920a786b025cb555461b6"} Nov 22 04:55:06 crc kubenswrapper[4818]: I1122 04:55:06.493937 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vcnvz" event={"ID":"603888a5-b8ab-4378-ab59-d044221e403c","Type":"ContainerStarted","Data":"24ca2a673e33c0e18266821325101a398eebcafa48e13e31e63b8014292f2b1c"} Nov 22 04:55:06 crc kubenswrapper[4818]: I1122 04:55:06.532056 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k4j4l" podStartSLOduration=2.476956097 podStartE2EDuration="5.532042571s" podCreationTimestamp="2025-11-22 04:55:01 +0000 UTC" firstStartedPulling="2025-11-22 04:55:02.453847779 +0000 UTC m=+455.028264306" lastFinishedPulling="2025-11-22 04:55:05.508934253 +0000 UTC m=+458.083350780" observedRunningTime="2025-11-22 04:55:06.529035278 +0000 UTC m=+459.103451805" watchObservedRunningTime="2025-11-22 04:55:06.532042571 +0000 UTC m=+459.106459108" Nov 22 04:55:06 crc kubenswrapper[4818]: I1122 04:55:06.551401 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vcnvz" podStartSLOduration=2.753000409 podStartE2EDuration="5.551386557s" podCreationTimestamp="2025-11-22 04:55:01 +0000 UTC" firstStartedPulling="2025-11-22 04:55:02.448760268 +0000 UTC m=+455.023176795" lastFinishedPulling="2025-11-22 04:55:05.247146416 +0000 UTC m=+457.821562943" observedRunningTime="2025-11-22 04:55:06.548478836 +0000 UTC m=+459.122895363" watchObservedRunningTime="2025-11-22 04:55:06.551386557 +0000 UTC m=+459.125803084" Nov 22 04:55:07 crc kubenswrapper[4818]: I1122 04:55:07.507951 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rcbjp" event={"ID":"d83d4bfc-8842-4d00-96d9-779a89080441","Type":"ContainerStarted","Data":"24693831385ed009c54ed2584a8ccc84ae0aee541e228060017dae22e80bcf99"} Nov 22 04:55:07 crc kubenswrapper[4818]: I1122 04:55:07.511713 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8lq9" event={"ID":"759a3ee8-f990-4960-bc61-9893ca8cbc6f","Type":"ContainerStarted","Data":"db980ced90ef2b1483f649e1a1dfb451daa6bd6648d3c4a398f038ebdd4c6f9c"} Nov 22 04:55:07 crc kubenswrapper[4818]: I1122 04:55:07.526312 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rcbjp" podStartSLOduration=2.067963807 podStartE2EDuration="4.526292762s" podCreationTimestamp="2025-11-22 04:55:03 +0000 UTC" firstStartedPulling="2025-11-22 04:55:04.471356373 +0000 UTC m=+457.045772900" lastFinishedPulling="2025-11-22 04:55:06.929685308 +0000 UTC m=+459.504101855" observedRunningTime="2025-11-22 04:55:07.523629839 +0000 UTC m=+460.098046366" watchObservedRunningTime="2025-11-22 04:55:07.526292762 +0000 UTC m=+460.100709289" Nov 22 04:55:07 crc kubenswrapper[4818]: I1122 04:55:07.544741 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h8lq9" podStartSLOduration=2.050440712 podStartE2EDuration="4.544719342s" podCreationTimestamp="2025-11-22 04:55:03 +0000 UTC" firstStartedPulling="2025-11-22 04:55:04.471297621 +0000 UTC m=+457.045714138" lastFinishedPulling="2025-11-22 04:55:06.965576241 +0000 UTC m=+459.539992768" observedRunningTime="2025-11-22 04:55:07.544544128 +0000 UTC m=+460.118960665" watchObservedRunningTime="2025-11-22 04:55:07.544719342 +0000 UTC m=+460.119135869" Nov 22 04:55:11 crc kubenswrapper[4818]: I1122 04:55:11.402712 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vcnvz" Nov 22 04:55:11 crc kubenswrapper[4818]: I1122 04:55:11.403444 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vcnvz" Nov 22 04:55:11 crc kubenswrapper[4818]: I1122 04:55:11.473235 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vcnvz" Nov 22 04:55:11 crc kubenswrapper[4818]: I1122 04:55:11.536358 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k4j4l" Nov 22 04:55:11 crc kubenswrapper[4818]: I1122 04:55:11.536430 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k4j4l" Nov 22 04:55:11 crc kubenswrapper[4818]: I1122 04:55:11.574186 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vcnvz" Nov 22 04:55:11 crc kubenswrapper[4818]: I1122 04:55:11.590438 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k4j4l" Nov 22 04:55:12 crc kubenswrapper[4818]: I1122 04:55:12.574744 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k4j4l" Nov 22 04:55:13 crc kubenswrapper[4818]: I1122 04:55:13.748116 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rcbjp" Nov 22 04:55:13 crc kubenswrapper[4818]: I1122 04:55:13.748576 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rcbjp" Nov 22 04:55:13 crc kubenswrapper[4818]: I1122 04:55:13.795687 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rcbjp" Nov 22 04:55:13 crc kubenswrapper[4818]: I1122 04:55:13.960824 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h8lq9" Nov 22 04:55:13 crc kubenswrapper[4818]: I1122 04:55:13.960910 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h8lq9" Nov 22 04:55:14 crc kubenswrapper[4818]: I1122 04:55:14.021307 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h8lq9" Nov 22 04:55:14 crc kubenswrapper[4818]: I1122 04:55:14.591636 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h8lq9" Nov 22 04:55:14 crc kubenswrapper[4818]: I1122 04:55:14.618006 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rcbjp" Nov 22 04:55:21 crc kubenswrapper[4818]: I1122 04:55:21.264277 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 04:55:21 crc kubenswrapper[4818]: I1122 04:55:21.264666 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 04:55:21 crc kubenswrapper[4818]: I1122 04:55:21.264738 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 04:55:21 crc kubenswrapper[4818]: I1122 04:55:21.265634 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"57a015280b24f5d226048157fcf5b764a30fd55452acd0f6fc61ec5f33defeca"} pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 04:55:21 crc kubenswrapper[4818]: I1122 04:55:21.265741 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" containerID="cri-o://57a015280b24f5d226048157fcf5b764a30fd55452acd0f6fc61ec5f33defeca" gracePeriod=600 Nov 22 04:55:21 crc kubenswrapper[4818]: I1122 04:55:21.591973 4818 generic.go:334] "Generic (PLEG): container finished" podID="0971c9ec-29ce-44eb-9875-ad4664143305" containerID="57a015280b24f5d226048157fcf5b764a30fd55452acd0f6fc61ec5f33defeca" exitCode=0 Nov 22 04:55:21 crc kubenswrapper[4818]: I1122 04:55:21.592099 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerDied","Data":"57a015280b24f5d226048157fcf5b764a30fd55452acd0f6fc61ec5f33defeca"} Nov 22 04:55:21 crc kubenswrapper[4818]: I1122 04:55:21.592430 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"86d6c8c4185dff63c83f0e08d741aef33efa784f30aa076579d91d5ab82bea2b"} Nov 22 04:55:21 crc kubenswrapper[4818]: I1122 04:55:21.592467 4818 scope.go:117] "RemoveContainer" containerID="2891adcaa3fb8ac3d149721705288966c034a7839316259cfdcdacfcad2b1863" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.577027 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-z5pms"] Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.579106 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.613833 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-z5pms"] Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.761787 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac79364e-e82a-472b-bd52-53949de37127-trusted-ca\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.761830 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxfwz\" (UniqueName: \"kubernetes.io/projected/ac79364e-e82a-472b-bd52-53949de37127-kube-api-access-sxfwz\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.761859 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ac79364e-e82a-472b-bd52-53949de37127-installation-pull-secrets\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.761877 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ac79364e-e82a-472b-bd52-53949de37127-bound-sa-token\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.761892 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ac79364e-e82a-472b-bd52-53949de37127-registry-certificates\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.761937 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ac79364e-e82a-472b-bd52-53949de37127-ca-trust-extracted\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.761958 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ac79364e-e82a-472b-bd52-53949de37127-registry-tls\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.761982 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.781486 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.863617 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac79364e-e82a-472b-bd52-53949de37127-trusted-ca\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.863678 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxfwz\" (UniqueName: \"kubernetes.io/projected/ac79364e-e82a-472b-bd52-53949de37127-kube-api-access-sxfwz\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.863710 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ac79364e-e82a-472b-bd52-53949de37127-installation-pull-secrets\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.863738 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ac79364e-e82a-472b-bd52-53949de37127-registry-certificates\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.863769 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ac79364e-e82a-472b-bd52-53949de37127-bound-sa-token\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.863831 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ac79364e-e82a-472b-bd52-53949de37127-ca-trust-extracted\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.863872 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ac79364e-e82a-472b-bd52-53949de37127-registry-tls\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.864368 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ac79364e-e82a-472b-bd52-53949de37127-ca-trust-extracted\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.865021 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ac79364e-e82a-472b-bd52-53949de37127-registry-certificates\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.865336 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac79364e-e82a-472b-bd52-53949de37127-trusted-ca\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.870118 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ac79364e-e82a-472b-bd52-53949de37127-installation-pull-secrets\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.870241 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ac79364e-e82a-472b-bd52-53949de37127-registry-tls\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.880554 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxfwz\" (UniqueName: \"kubernetes.io/projected/ac79364e-e82a-472b-bd52-53949de37127-kube-api-access-sxfwz\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.880663 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ac79364e-e82a-472b-bd52-53949de37127-bound-sa-token\") pod \"image-registry-66df7c8f76-z5pms\" (UID: \"ac79364e-e82a-472b-bd52-53949de37127\") " pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:32 crc kubenswrapper[4818]: I1122 04:55:32.904496 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:33 crc kubenswrapper[4818]: I1122 04:55:33.104987 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-z5pms"] Nov 22 04:55:33 crc kubenswrapper[4818]: I1122 04:55:33.664411 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" event={"ID":"ac79364e-e82a-472b-bd52-53949de37127","Type":"ContainerStarted","Data":"d63d6b47df79d348cb93db892c373bda2fdda8f3bd67015f07ce41cbf1098e38"} Nov 22 04:55:54 crc kubenswrapper[4818]: I1122 04:55:54.806442 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" event={"ID":"ac79364e-e82a-472b-bd52-53949de37127","Type":"ContainerStarted","Data":"b42060eae0333ebf1dc6ecd8eddcd3c359dce952081eeccca13396ab8d56099a"} Nov 22 04:55:55 crc kubenswrapper[4818]: I1122 04:55:55.813013 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:55:55 crc kubenswrapper[4818]: I1122 04:55:55.843326 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" podStartSLOduration=23.843306401 podStartE2EDuration="23.843306401s" podCreationTimestamp="2025-11-22 04:55:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 04:55:55.840508844 +0000 UTC m=+508.414925411" watchObservedRunningTime="2025-11-22 04:55:55.843306401 +0000 UTC m=+508.417722938" Nov 22 04:56:12 crc kubenswrapper[4818]: I1122 04:56:12.913565 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-z5pms" Nov 22 04:56:12 crc kubenswrapper[4818]: I1122 04:56:12.997427 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cd4qt"] Nov 22 04:56:38 crc kubenswrapper[4818]: I1122 04:56:38.056748 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" podUID="677aa266-f94d-498a-8034-1205889c4958" containerName="registry" containerID="cri-o://3d29507ef31e0257cdef51333c6adb2b79403040b5a6dda3dc6df109fe9fdb41" gracePeriod=30 Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.051146 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.097795 4818 generic.go:334] "Generic (PLEG): container finished" podID="677aa266-f94d-498a-8034-1205889c4958" containerID="3d29507ef31e0257cdef51333c6adb2b79403040b5a6dda3dc6df109fe9fdb41" exitCode=0 Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.097830 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" event={"ID":"677aa266-f94d-498a-8034-1205889c4958","Type":"ContainerDied","Data":"3d29507ef31e0257cdef51333c6adb2b79403040b5a6dda3dc6df109fe9fdb41"} Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.097853 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.097872 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cd4qt" event={"ID":"677aa266-f94d-498a-8034-1205889c4958","Type":"ContainerDied","Data":"90651c0e4b27fe0f9c8b8aa42ec12e3e28d4c4d2485a6455f7d7d546e5bc133a"} Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.097919 4818 scope.go:117] "RemoveContainer" containerID="3d29507ef31e0257cdef51333c6adb2b79403040b5a6dda3dc6df109fe9fdb41" Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.115828 4818 scope.go:117] "RemoveContainer" containerID="3d29507ef31e0257cdef51333c6adb2b79403040b5a6dda3dc6df109fe9fdb41" Nov 22 04:56:39 crc kubenswrapper[4818]: E1122 04:56:39.116212 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d29507ef31e0257cdef51333c6adb2b79403040b5a6dda3dc6df109fe9fdb41\": container with ID starting with 3d29507ef31e0257cdef51333c6adb2b79403040b5a6dda3dc6df109fe9fdb41 not found: ID does not exist" containerID="3d29507ef31e0257cdef51333c6adb2b79403040b5a6dda3dc6df109fe9fdb41" Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.116249 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d29507ef31e0257cdef51333c6adb2b79403040b5a6dda3dc6df109fe9fdb41"} err="failed to get container status \"3d29507ef31e0257cdef51333c6adb2b79403040b5a6dda3dc6df109fe9fdb41\": rpc error: code = NotFound desc = could not find container \"3d29507ef31e0257cdef51333c6adb2b79403040b5a6dda3dc6df109fe9fdb41\": container with ID starting with 3d29507ef31e0257cdef51333c6adb2b79403040b5a6dda3dc6df109fe9fdb41 not found: ID does not exist" Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.223108 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/677aa266-f94d-498a-8034-1205889c4958-installation-pull-secrets\") pod \"677aa266-f94d-498a-8034-1205889c4958\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.223283 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/677aa266-f94d-498a-8034-1205889c4958-registry-certificates\") pod \"677aa266-f94d-498a-8034-1205889c4958\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.223331 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/677aa266-f94d-498a-8034-1205889c4958-ca-trust-extracted\") pod \"677aa266-f94d-498a-8034-1205889c4958\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.223361 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/677aa266-f94d-498a-8034-1205889c4958-registry-tls\") pod \"677aa266-f94d-498a-8034-1205889c4958\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.223391 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/677aa266-f94d-498a-8034-1205889c4958-bound-sa-token\") pod \"677aa266-f94d-498a-8034-1205889c4958\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.223564 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"677aa266-f94d-498a-8034-1205889c4958\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.223610 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/677aa266-f94d-498a-8034-1205889c4958-trusted-ca\") pod \"677aa266-f94d-498a-8034-1205889c4958\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.223659 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5fdv\" (UniqueName: \"kubernetes.io/projected/677aa266-f94d-498a-8034-1205889c4958-kube-api-access-n5fdv\") pod \"677aa266-f94d-498a-8034-1205889c4958\" (UID: \"677aa266-f94d-498a-8034-1205889c4958\") " Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.224973 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/677aa266-f94d-498a-8034-1205889c4958-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "677aa266-f94d-498a-8034-1205889c4958" (UID: "677aa266-f94d-498a-8034-1205889c4958"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.226410 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/677aa266-f94d-498a-8034-1205889c4958-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "677aa266-f94d-498a-8034-1205889c4958" (UID: "677aa266-f94d-498a-8034-1205889c4958"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.230683 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/677aa266-f94d-498a-8034-1205889c4958-kube-api-access-n5fdv" (OuterVolumeSpecName: "kube-api-access-n5fdv") pod "677aa266-f94d-498a-8034-1205889c4958" (UID: "677aa266-f94d-498a-8034-1205889c4958"). InnerVolumeSpecName "kube-api-access-n5fdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.230839 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/677aa266-f94d-498a-8034-1205889c4958-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "677aa266-f94d-498a-8034-1205889c4958" (UID: "677aa266-f94d-498a-8034-1205889c4958"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.231086 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/677aa266-f94d-498a-8034-1205889c4958-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "677aa266-f94d-498a-8034-1205889c4958" (UID: "677aa266-f94d-498a-8034-1205889c4958"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.235498 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/677aa266-f94d-498a-8034-1205889c4958-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "677aa266-f94d-498a-8034-1205889c4958" (UID: "677aa266-f94d-498a-8034-1205889c4958"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.245635 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "677aa266-f94d-498a-8034-1205889c4958" (UID: "677aa266-f94d-498a-8034-1205889c4958"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.257768 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/677aa266-f94d-498a-8034-1205889c4958-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "677aa266-f94d-498a-8034-1205889c4958" (UID: "677aa266-f94d-498a-8034-1205889c4958"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.325613 4818 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/677aa266-f94d-498a-8034-1205889c4958-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.325659 4818 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/677aa266-f94d-498a-8034-1205889c4958-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.325679 4818 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/677aa266-f94d-498a-8034-1205889c4958-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.325699 4818 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/677aa266-f94d-498a-8034-1205889c4958-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.325716 4818 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/677aa266-f94d-498a-8034-1205889c4958-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.325732 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5fdv\" (UniqueName: \"kubernetes.io/projected/677aa266-f94d-498a-8034-1205889c4958-kube-api-access-n5fdv\") on node \"crc\" DevicePath \"\"" Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.325750 4818 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/677aa266-f94d-498a-8034-1205889c4958-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.442218 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cd4qt"] Nov 22 04:56:39 crc kubenswrapper[4818]: I1122 04:56:39.448014 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cd4qt"] Nov 22 04:56:40 crc kubenswrapper[4818]: I1122 04:56:40.299982 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="677aa266-f94d-498a-8034-1205889c4958" path="/var/lib/kubelet/pods/677aa266-f94d-498a-8034-1205889c4958/volumes" Nov 22 04:57:21 crc kubenswrapper[4818]: I1122 04:57:21.265103 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 04:57:21 crc kubenswrapper[4818]: I1122 04:57:21.265910 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 04:57:51 crc kubenswrapper[4818]: I1122 04:57:51.264993 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 04:57:51 crc kubenswrapper[4818]: I1122 04:57:51.265994 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 04:58:21 crc kubenswrapper[4818]: I1122 04:58:21.264615 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 04:58:21 crc kubenswrapper[4818]: I1122 04:58:21.265152 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 04:58:21 crc kubenswrapper[4818]: I1122 04:58:21.265242 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 04:58:21 crc kubenswrapper[4818]: I1122 04:58:21.266326 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"86d6c8c4185dff63c83f0e08d741aef33efa784f30aa076579d91d5ab82bea2b"} pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 04:58:21 crc kubenswrapper[4818]: I1122 04:58:21.266458 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" containerID="cri-o://86d6c8c4185dff63c83f0e08d741aef33efa784f30aa076579d91d5ab82bea2b" gracePeriod=600 Nov 22 04:58:21 crc kubenswrapper[4818]: I1122 04:58:21.763126 4818 generic.go:334] "Generic (PLEG): container finished" podID="0971c9ec-29ce-44eb-9875-ad4664143305" containerID="86d6c8c4185dff63c83f0e08d741aef33efa784f30aa076579d91d5ab82bea2b" exitCode=0 Nov 22 04:58:21 crc kubenswrapper[4818]: I1122 04:58:21.763221 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerDied","Data":"86d6c8c4185dff63c83f0e08d741aef33efa784f30aa076579d91d5ab82bea2b"} Nov 22 04:58:21 crc kubenswrapper[4818]: I1122 04:58:21.763454 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"0c9825f8f0cd7e721d11fe80d62d92bcabce6091d17bb472097983d66a29cf6a"} Nov 22 04:58:21 crc kubenswrapper[4818]: I1122 04:58:21.763490 4818 scope.go:117] "RemoveContainer" containerID="57a015280b24f5d226048157fcf5b764a30fd55452acd0f6fc61ec5f33defeca" Nov 22 05:00:00 crc kubenswrapper[4818]: I1122 05:00:00.152190 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq"] Nov 22 05:00:00 crc kubenswrapper[4818]: E1122 05:00:00.153161 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="677aa266-f94d-498a-8034-1205889c4958" containerName="registry" Nov 22 05:00:00 crc kubenswrapper[4818]: I1122 05:00:00.153189 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="677aa266-f94d-498a-8034-1205889c4958" containerName="registry" Nov 22 05:00:00 crc kubenswrapper[4818]: I1122 05:00:00.153412 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="677aa266-f94d-498a-8034-1205889c4958" containerName="registry" Nov 22 05:00:00 crc kubenswrapper[4818]: I1122 05:00:00.154032 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq" Nov 22 05:00:00 crc kubenswrapper[4818]: I1122 05:00:00.160611 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 05:00:00 crc kubenswrapper[4818]: I1122 05:00:00.167755 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq"] Nov 22 05:00:00 crc kubenswrapper[4818]: I1122 05:00:00.168302 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 05:00:00 crc kubenswrapper[4818]: I1122 05:00:00.190169 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd26x\" (UniqueName: \"kubernetes.io/projected/176f1072-37ec-476e-8bcc-eda2271dd884-kube-api-access-gd26x\") pod \"collect-profiles-29396460-hb4dq\" (UID: \"176f1072-37ec-476e-8bcc-eda2271dd884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq" Nov 22 05:00:00 crc kubenswrapper[4818]: I1122 05:00:00.190233 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/176f1072-37ec-476e-8bcc-eda2271dd884-secret-volume\") pod \"collect-profiles-29396460-hb4dq\" (UID: \"176f1072-37ec-476e-8bcc-eda2271dd884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq" Nov 22 05:00:00 crc kubenswrapper[4818]: I1122 05:00:00.190338 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/176f1072-37ec-476e-8bcc-eda2271dd884-config-volume\") pod \"collect-profiles-29396460-hb4dq\" (UID: \"176f1072-37ec-476e-8bcc-eda2271dd884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq" Nov 22 05:00:00 crc kubenswrapper[4818]: I1122 05:00:00.290775 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd26x\" (UniqueName: \"kubernetes.io/projected/176f1072-37ec-476e-8bcc-eda2271dd884-kube-api-access-gd26x\") pod \"collect-profiles-29396460-hb4dq\" (UID: \"176f1072-37ec-476e-8bcc-eda2271dd884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq" Nov 22 05:00:00 crc kubenswrapper[4818]: I1122 05:00:00.290815 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/176f1072-37ec-476e-8bcc-eda2271dd884-secret-volume\") pod \"collect-profiles-29396460-hb4dq\" (UID: \"176f1072-37ec-476e-8bcc-eda2271dd884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq" Nov 22 05:00:00 crc kubenswrapper[4818]: I1122 05:00:00.290860 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/176f1072-37ec-476e-8bcc-eda2271dd884-config-volume\") pod \"collect-profiles-29396460-hb4dq\" (UID: \"176f1072-37ec-476e-8bcc-eda2271dd884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq" Nov 22 05:00:00 crc kubenswrapper[4818]: I1122 05:00:00.291711 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/176f1072-37ec-476e-8bcc-eda2271dd884-config-volume\") pod \"collect-profiles-29396460-hb4dq\" (UID: \"176f1072-37ec-476e-8bcc-eda2271dd884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq" Nov 22 05:00:00 crc kubenswrapper[4818]: I1122 05:00:00.296548 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/176f1072-37ec-476e-8bcc-eda2271dd884-secret-volume\") pod \"collect-profiles-29396460-hb4dq\" (UID: \"176f1072-37ec-476e-8bcc-eda2271dd884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq" Nov 22 05:00:00 crc kubenswrapper[4818]: I1122 05:00:00.305207 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd26x\" (UniqueName: \"kubernetes.io/projected/176f1072-37ec-476e-8bcc-eda2271dd884-kube-api-access-gd26x\") pod \"collect-profiles-29396460-hb4dq\" (UID: \"176f1072-37ec-476e-8bcc-eda2271dd884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq" Nov 22 05:00:00 crc kubenswrapper[4818]: I1122 05:00:00.480895 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq" Nov 22 05:00:00 crc kubenswrapper[4818]: I1122 05:00:00.696693 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq"] Nov 22 05:00:01 crc kubenswrapper[4818]: I1122 05:00:01.418748 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq" event={"ID":"176f1072-37ec-476e-8bcc-eda2271dd884","Type":"ContainerStarted","Data":"2ff61c4f9d6d40fc8a93a3eedf71ee13fa18a916b73695691c0b347cc8abee1a"} Nov 22 05:00:01 crc kubenswrapper[4818]: I1122 05:00:01.419929 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq" event={"ID":"176f1072-37ec-476e-8bcc-eda2271dd884","Type":"ContainerStarted","Data":"4f9f830bbd9837453081cfafc2654898ae0d5f830eb002a416b52f5c6d3c4cbc"} Nov 22 05:00:01 crc kubenswrapper[4818]: I1122 05:00:01.438813 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq" podStartSLOduration=1.438794257 podStartE2EDuration="1.438794257s" podCreationTimestamp="2025-11-22 05:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:00:01.435605383 +0000 UTC m=+754.010021930" watchObservedRunningTime="2025-11-22 05:00:01.438794257 +0000 UTC m=+754.013210784" Nov 22 05:00:02 crc kubenswrapper[4818]: I1122 05:00:02.426427 4818 generic.go:334] "Generic (PLEG): container finished" podID="176f1072-37ec-476e-8bcc-eda2271dd884" containerID="2ff61c4f9d6d40fc8a93a3eedf71ee13fa18a916b73695691c0b347cc8abee1a" exitCode=0 Nov 22 05:00:02 crc kubenswrapper[4818]: I1122 05:00:02.427364 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq" event={"ID":"176f1072-37ec-476e-8bcc-eda2271dd884","Type":"ContainerDied","Data":"2ff61c4f9d6d40fc8a93a3eedf71ee13fa18a916b73695691c0b347cc8abee1a"} Nov 22 05:00:03 crc kubenswrapper[4818]: I1122 05:00:03.732297 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq" Nov 22 05:00:03 crc kubenswrapper[4818]: I1122 05:00:03.745839 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/176f1072-37ec-476e-8bcc-eda2271dd884-config-volume\") pod \"176f1072-37ec-476e-8bcc-eda2271dd884\" (UID: \"176f1072-37ec-476e-8bcc-eda2271dd884\") " Nov 22 05:00:03 crc kubenswrapper[4818]: I1122 05:00:03.745918 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gd26x\" (UniqueName: \"kubernetes.io/projected/176f1072-37ec-476e-8bcc-eda2271dd884-kube-api-access-gd26x\") pod \"176f1072-37ec-476e-8bcc-eda2271dd884\" (UID: \"176f1072-37ec-476e-8bcc-eda2271dd884\") " Nov 22 05:00:03 crc kubenswrapper[4818]: I1122 05:00:03.746021 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/176f1072-37ec-476e-8bcc-eda2271dd884-secret-volume\") pod \"176f1072-37ec-476e-8bcc-eda2271dd884\" (UID: \"176f1072-37ec-476e-8bcc-eda2271dd884\") " Nov 22 05:00:03 crc kubenswrapper[4818]: I1122 05:00:03.746567 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/176f1072-37ec-476e-8bcc-eda2271dd884-config-volume" (OuterVolumeSpecName: "config-volume") pod "176f1072-37ec-476e-8bcc-eda2271dd884" (UID: "176f1072-37ec-476e-8bcc-eda2271dd884"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:00:03 crc kubenswrapper[4818]: I1122 05:00:03.755090 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/176f1072-37ec-476e-8bcc-eda2271dd884-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "176f1072-37ec-476e-8bcc-eda2271dd884" (UID: "176f1072-37ec-476e-8bcc-eda2271dd884"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:00:03 crc kubenswrapper[4818]: I1122 05:00:03.755324 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/176f1072-37ec-476e-8bcc-eda2271dd884-kube-api-access-gd26x" (OuterVolumeSpecName: "kube-api-access-gd26x") pod "176f1072-37ec-476e-8bcc-eda2271dd884" (UID: "176f1072-37ec-476e-8bcc-eda2271dd884"). InnerVolumeSpecName "kube-api-access-gd26x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:00:03 crc kubenswrapper[4818]: I1122 05:00:03.848041 4818 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/176f1072-37ec-476e-8bcc-eda2271dd884-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 05:00:03 crc kubenswrapper[4818]: I1122 05:00:03.848072 4818 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/176f1072-37ec-476e-8bcc-eda2271dd884-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 05:00:03 crc kubenswrapper[4818]: I1122 05:00:03.848082 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gd26x\" (UniqueName: \"kubernetes.io/projected/176f1072-37ec-476e-8bcc-eda2271dd884-kube-api-access-gd26x\") on node \"crc\" DevicePath \"\"" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.086412 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tjh65"] Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.086925 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" podUID="d5484670-b8e2-40b9-a36b-3fb47bb89421" containerName="controller-manager" containerID="cri-o://726de57de72645b656493b838692c8024ebb101a30d9def67c7ffc96630fd79c" gracePeriod=30 Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.183198 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x"] Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.183459 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" podUID="353ed841-0d3b-46df-beec-1906f6d801fd" containerName="route-controller-manager" containerID="cri-o://2c302fc90978cceb46d9b02614009b42c6092cfb0211756cb5b368a9f782e821" gracePeriod=30 Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.405144 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.455477 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58r46\" (UniqueName: \"kubernetes.io/projected/d5484670-b8e2-40b9-a36b-3fb47bb89421-kube-api-access-58r46\") pod \"d5484670-b8e2-40b9-a36b-3fb47bb89421\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.455515 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5484670-b8e2-40b9-a36b-3fb47bb89421-serving-cert\") pod \"d5484670-b8e2-40b9-a36b-3fb47bb89421\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.455579 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-client-ca\") pod \"d5484670-b8e2-40b9-a36b-3fb47bb89421\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.455595 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-proxy-ca-bundles\") pod \"d5484670-b8e2-40b9-a36b-3fb47bb89421\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.455613 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-config\") pod \"d5484670-b8e2-40b9-a36b-3fb47bb89421\" (UID: \"d5484670-b8e2-40b9-a36b-3fb47bb89421\") " Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.456685 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-config" (OuterVolumeSpecName: "config") pod "d5484670-b8e2-40b9-a36b-3fb47bb89421" (UID: "d5484670-b8e2-40b9-a36b-3fb47bb89421"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.458032 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-client-ca" (OuterVolumeSpecName: "client-ca") pod "d5484670-b8e2-40b9-a36b-3fb47bb89421" (UID: "d5484670-b8e2-40b9-a36b-3fb47bb89421"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.458411 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "d5484670-b8e2-40b9-a36b-3fb47bb89421" (UID: "d5484670-b8e2-40b9-a36b-3fb47bb89421"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.465139 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5484670-b8e2-40b9-a36b-3fb47bb89421-kube-api-access-58r46" (OuterVolumeSpecName: "kube-api-access-58r46") pod "d5484670-b8e2-40b9-a36b-3fb47bb89421" (UID: "d5484670-b8e2-40b9-a36b-3fb47bb89421"). InnerVolumeSpecName "kube-api-access-58r46". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.465565 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5484670-b8e2-40b9-a36b-3fb47bb89421-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d5484670-b8e2-40b9-a36b-3fb47bb89421" (UID: "d5484670-b8e2-40b9-a36b-3fb47bb89421"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.479423 4818 generic.go:334] "Generic (PLEG): container finished" podID="d5484670-b8e2-40b9-a36b-3fb47bb89421" containerID="726de57de72645b656493b838692c8024ebb101a30d9def67c7ffc96630fd79c" exitCode=0 Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.479608 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.480339 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" event={"ID":"d5484670-b8e2-40b9-a36b-3fb47bb89421","Type":"ContainerDied","Data":"726de57de72645b656493b838692c8024ebb101a30d9def67c7ffc96630fd79c"} Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.480552 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tjh65" event={"ID":"d5484670-b8e2-40b9-a36b-3fb47bb89421","Type":"ContainerDied","Data":"fe11bd277dcae022d5866f9a2af875bb91519a95a25776307fc641ef2efa74f9"} Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.480766 4818 scope.go:117] "RemoveContainer" containerID="726de57de72645b656493b838692c8024ebb101a30d9def67c7ffc96630fd79c" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.487867 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.488715 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq" event={"ID":"176f1072-37ec-476e-8bcc-eda2271dd884","Type":"ContainerDied","Data":"4f9f830bbd9837453081cfafc2654898ae0d5f830eb002a416b52f5c6d3c4cbc"} Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.488753 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f9f830bbd9837453081cfafc2654898ae0d5f830eb002a416b52f5c6d3c4cbc" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.496554 4818 generic.go:334] "Generic (PLEG): container finished" podID="353ed841-0d3b-46df-beec-1906f6d801fd" containerID="2c302fc90978cceb46d9b02614009b42c6092cfb0211756cb5b368a9f782e821" exitCode=0 Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.496593 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" event={"ID":"353ed841-0d3b-46df-beec-1906f6d801fd","Type":"ContainerDied","Data":"2c302fc90978cceb46d9b02614009b42c6092cfb0211756cb5b368a9f782e821"} Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.521326 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tjh65"] Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.523889 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tjh65"] Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.528811 4818 scope.go:117] "RemoveContainer" containerID="726de57de72645b656493b838692c8024ebb101a30d9def67c7ffc96630fd79c" Nov 22 05:00:04 crc kubenswrapper[4818]: E1122 05:00:04.529551 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"726de57de72645b656493b838692c8024ebb101a30d9def67c7ffc96630fd79c\": container with ID starting with 726de57de72645b656493b838692c8024ebb101a30d9def67c7ffc96630fd79c not found: ID does not exist" containerID="726de57de72645b656493b838692c8024ebb101a30d9def67c7ffc96630fd79c" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.529581 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"726de57de72645b656493b838692c8024ebb101a30d9def67c7ffc96630fd79c"} err="failed to get container status \"726de57de72645b656493b838692c8024ebb101a30d9def67c7ffc96630fd79c\": rpc error: code = NotFound desc = could not find container \"726de57de72645b656493b838692c8024ebb101a30d9def67c7ffc96630fd79c\": container with ID starting with 726de57de72645b656493b838692c8024ebb101a30d9def67c7ffc96630fd79c not found: ID does not exist" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.557185 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.557229 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58r46\" (UniqueName: \"kubernetes.io/projected/d5484670-b8e2-40b9-a36b-3fb47bb89421-kube-api-access-58r46\") on node \"crc\" DevicePath \"\"" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.557285 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5484670-b8e2-40b9-a36b-3fb47bb89421-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.557311 4818 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-client-ca\") on node \"crc\" DevicePath \"\"" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.557334 4818 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d5484670-b8e2-40b9-a36b-3fb47bb89421-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.605115 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.658399 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqrct\" (UniqueName: \"kubernetes.io/projected/353ed841-0d3b-46df-beec-1906f6d801fd-kube-api-access-fqrct\") pod \"353ed841-0d3b-46df-beec-1906f6d801fd\" (UID: \"353ed841-0d3b-46df-beec-1906f6d801fd\") " Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.658494 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/353ed841-0d3b-46df-beec-1906f6d801fd-config\") pod \"353ed841-0d3b-46df-beec-1906f6d801fd\" (UID: \"353ed841-0d3b-46df-beec-1906f6d801fd\") " Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.658589 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/353ed841-0d3b-46df-beec-1906f6d801fd-serving-cert\") pod \"353ed841-0d3b-46df-beec-1906f6d801fd\" (UID: \"353ed841-0d3b-46df-beec-1906f6d801fd\") " Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.658657 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/353ed841-0d3b-46df-beec-1906f6d801fd-client-ca\") pod \"353ed841-0d3b-46df-beec-1906f6d801fd\" (UID: \"353ed841-0d3b-46df-beec-1906f6d801fd\") " Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.659366 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/353ed841-0d3b-46df-beec-1906f6d801fd-config" (OuterVolumeSpecName: "config") pod "353ed841-0d3b-46df-beec-1906f6d801fd" (UID: "353ed841-0d3b-46df-beec-1906f6d801fd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.659501 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/353ed841-0d3b-46df-beec-1906f6d801fd-client-ca" (OuterVolumeSpecName: "client-ca") pod "353ed841-0d3b-46df-beec-1906f6d801fd" (UID: "353ed841-0d3b-46df-beec-1906f6d801fd"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.662672 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/353ed841-0d3b-46df-beec-1906f6d801fd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "353ed841-0d3b-46df-beec-1906f6d801fd" (UID: "353ed841-0d3b-46df-beec-1906f6d801fd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.662941 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/353ed841-0d3b-46df-beec-1906f6d801fd-kube-api-access-fqrct" (OuterVolumeSpecName: "kube-api-access-fqrct") pod "353ed841-0d3b-46df-beec-1906f6d801fd" (UID: "353ed841-0d3b-46df-beec-1906f6d801fd"). InnerVolumeSpecName "kube-api-access-fqrct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.759747 4818 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/353ed841-0d3b-46df-beec-1906f6d801fd-client-ca\") on node \"crc\" DevicePath \"\"" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.759795 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqrct\" (UniqueName: \"kubernetes.io/projected/353ed841-0d3b-46df-beec-1906f6d801fd-kube-api-access-fqrct\") on node \"crc\" DevicePath \"\"" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.759814 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/353ed841-0d3b-46df-beec-1906f6d801fd-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:00:04 crc kubenswrapper[4818]: I1122 05:00:04.759831 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/353ed841-0d3b-46df-beec-1906f6d801fd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.504729 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" event={"ID":"353ed841-0d3b-46df-beec-1906f6d801fd","Type":"ContainerDied","Data":"54cea90252e2fa553613928256095d4621431b359ca2c50a528b05479c915594"} Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.504759 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.504791 4818 scope.go:117] "RemoveContainer" containerID="2c302fc90978cceb46d9b02614009b42c6092cfb0211756cb5b368a9f782e821" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.551789 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x"] Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.558425 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4lz4x"] Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.707928 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8"] Nov 22 05:00:05 crc kubenswrapper[4818]: E1122 05:00:05.708230 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5484670-b8e2-40b9-a36b-3fb47bb89421" containerName="controller-manager" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.708287 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5484670-b8e2-40b9-a36b-3fb47bb89421" containerName="controller-manager" Nov 22 05:00:05 crc kubenswrapper[4818]: E1122 05:00:05.708309 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="176f1072-37ec-476e-8bcc-eda2271dd884" containerName="collect-profiles" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.708323 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="176f1072-37ec-476e-8bcc-eda2271dd884" containerName="collect-profiles" Nov 22 05:00:05 crc kubenswrapper[4818]: E1122 05:00:05.708352 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="353ed841-0d3b-46df-beec-1906f6d801fd" containerName="route-controller-manager" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.708368 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="353ed841-0d3b-46df-beec-1906f6d801fd" containerName="route-controller-manager" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.708531 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="353ed841-0d3b-46df-beec-1906f6d801fd" containerName="route-controller-manager" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.708558 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5484670-b8e2-40b9-a36b-3fb47bb89421" containerName="controller-manager" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.708588 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="176f1072-37ec-476e-8bcc-eda2271dd884" containerName="collect-profiles" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.709185 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.714244 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.714355 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.716037 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.716764 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.716766 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.718215 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk"] Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.719021 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.720415 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.723719 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.723949 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.724467 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.724518 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.725632 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.730796 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8"] Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.732411 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.733576 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.741507 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk"] Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.773305 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e7c63aa-6e42-4f75-a595-69df5005d036-config\") pod \"route-controller-manager-774dd8f765-4l6hk\" (UID: \"2e7c63aa-6e42-4f75-a595-69df5005d036\") " pod="openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.773374 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adeee5af-b824-4034-acf1-8cd7a4931b5a-config\") pod \"controller-manager-7cb6cd57bb-rz7c8\" (UID: \"adeee5af-b824-4034-acf1-8cd7a4931b5a\") " pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.773429 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/adeee5af-b824-4034-acf1-8cd7a4931b5a-proxy-ca-bundles\") pod \"controller-manager-7cb6cd57bb-rz7c8\" (UID: \"adeee5af-b824-4034-acf1-8cd7a4931b5a\") " pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.773470 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2e7c63aa-6e42-4f75-a595-69df5005d036-client-ca\") pod \"route-controller-manager-774dd8f765-4l6hk\" (UID: \"2e7c63aa-6e42-4f75-a595-69df5005d036\") " pod="openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.773542 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6cxz\" (UniqueName: \"kubernetes.io/projected/2e7c63aa-6e42-4f75-a595-69df5005d036-kube-api-access-k6cxz\") pod \"route-controller-manager-774dd8f765-4l6hk\" (UID: \"2e7c63aa-6e42-4f75-a595-69df5005d036\") " pod="openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.773572 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e7c63aa-6e42-4f75-a595-69df5005d036-serving-cert\") pod \"route-controller-manager-774dd8f765-4l6hk\" (UID: \"2e7c63aa-6e42-4f75-a595-69df5005d036\") " pod="openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.773597 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/adeee5af-b824-4034-acf1-8cd7a4931b5a-client-ca\") pod \"controller-manager-7cb6cd57bb-rz7c8\" (UID: \"adeee5af-b824-4034-acf1-8cd7a4931b5a\") " pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.773624 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftwd4\" (UniqueName: \"kubernetes.io/projected/adeee5af-b824-4034-acf1-8cd7a4931b5a-kube-api-access-ftwd4\") pod \"controller-manager-7cb6cd57bb-rz7c8\" (UID: \"adeee5af-b824-4034-acf1-8cd7a4931b5a\") " pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.773728 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adeee5af-b824-4034-acf1-8cd7a4931b5a-serving-cert\") pod \"controller-manager-7cb6cd57bb-rz7c8\" (UID: \"adeee5af-b824-4034-acf1-8cd7a4931b5a\") " pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.874808 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftwd4\" (UniqueName: \"kubernetes.io/projected/adeee5af-b824-4034-acf1-8cd7a4931b5a-kube-api-access-ftwd4\") pod \"controller-manager-7cb6cd57bb-rz7c8\" (UID: \"adeee5af-b824-4034-acf1-8cd7a4931b5a\") " pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.875084 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adeee5af-b824-4034-acf1-8cd7a4931b5a-serving-cert\") pod \"controller-manager-7cb6cd57bb-rz7c8\" (UID: \"adeee5af-b824-4034-acf1-8cd7a4931b5a\") " pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.875207 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e7c63aa-6e42-4f75-a595-69df5005d036-config\") pod \"route-controller-manager-774dd8f765-4l6hk\" (UID: \"2e7c63aa-6e42-4f75-a595-69df5005d036\") " pod="openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.875348 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adeee5af-b824-4034-acf1-8cd7a4931b5a-config\") pod \"controller-manager-7cb6cd57bb-rz7c8\" (UID: \"adeee5af-b824-4034-acf1-8cd7a4931b5a\") " pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.875472 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/adeee5af-b824-4034-acf1-8cd7a4931b5a-proxy-ca-bundles\") pod \"controller-manager-7cb6cd57bb-rz7c8\" (UID: \"adeee5af-b824-4034-acf1-8cd7a4931b5a\") " pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.875575 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2e7c63aa-6e42-4f75-a595-69df5005d036-client-ca\") pod \"route-controller-manager-774dd8f765-4l6hk\" (UID: \"2e7c63aa-6e42-4f75-a595-69df5005d036\") " pod="openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.875683 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6cxz\" (UniqueName: \"kubernetes.io/projected/2e7c63aa-6e42-4f75-a595-69df5005d036-kube-api-access-k6cxz\") pod \"route-controller-manager-774dd8f765-4l6hk\" (UID: \"2e7c63aa-6e42-4f75-a595-69df5005d036\") " pod="openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.875786 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e7c63aa-6e42-4f75-a595-69df5005d036-serving-cert\") pod \"route-controller-manager-774dd8f765-4l6hk\" (UID: \"2e7c63aa-6e42-4f75-a595-69df5005d036\") " pod="openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.875897 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/adeee5af-b824-4034-acf1-8cd7a4931b5a-client-ca\") pod \"controller-manager-7cb6cd57bb-rz7c8\" (UID: \"adeee5af-b824-4034-acf1-8cd7a4931b5a\") " pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.876563 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e7c63aa-6e42-4f75-a595-69df5005d036-config\") pod \"route-controller-manager-774dd8f765-4l6hk\" (UID: \"2e7c63aa-6e42-4f75-a595-69df5005d036\") " pod="openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.876653 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adeee5af-b824-4034-acf1-8cd7a4931b5a-config\") pod \"controller-manager-7cb6cd57bb-rz7c8\" (UID: \"adeee5af-b824-4034-acf1-8cd7a4931b5a\") " pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.876721 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/adeee5af-b824-4034-acf1-8cd7a4931b5a-client-ca\") pod \"controller-manager-7cb6cd57bb-rz7c8\" (UID: \"adeee5af-b824-4034-acf1-8cd7a4931b5a\") " pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.876769 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/adeee5af-b824-4034-acf1-8cd7a4931b5a-proxy-ca-bundles\") pod \"controller-manager-7cb6cd57bb-rz7c8\" (UID: \"adeee5af-b824-4034-acf1-8cd7a4931b5a\") " pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.877278 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2e7c63aa-6e42-4f75-a595-69df5005d036-client-ca\") pod \"route-controller-manager-774dd8f765-4l6hk\" (UID: \"2e7c63aa-6e42-4f75-a595-69df5005d036\") " pod="openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.882910 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e7c63aa-6e42-4f75-a595-69df5005d036-serving-cert\") pod \"route-controller-manager-774dd8f765-4l6hk\" (UID: \"2e7c63aa-6e42-4f75-a595-69df5005d036\") " pod="openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.882936 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adeee5af-b824-4034-acf1-8cd7a4931b5a-serving-cert\") pod \"controller-manager-7cb6cd57bb-rz7c8\" (UID: \"adeee5af-b824-4034-acf1-8cd7a4931b5a\") " pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.890971 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftwd4\" (UniqueName: \"kubernetes.io/projected/adeee5af-b824-4034-acf1-8cd7a4931b5a-kube-api-access-ftwd4\") pod \"controller-manager-7cb6cd57bb-rz7c8\" (UID: \"adeee5af-b824-4034-acf1-8cd7a4931b5a\") " pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" Nov 22 05:00:05 crc kubenswrapper[4818]: I1122 05:00:05.890984 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6cxz\" (UniqueName: \"kubernetes.io/projected/2e7c63aa-6e42-4f75-a595-69df5005d036-kube-api-access-k6cxz\") pod \"route-controller-manager-774dd8f765-4l6hk\" (UID: \"2e7c63aa-6e42-4f75-a595-69df5005d036\") " pod="openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk" Nov 22 05:00:06 crc kubenswrapper[4818]: I1122 05:00:06.026381 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" Nov 22 05:00:06 crc kubenswrapper[4818]: I1122 05:00:06.039885 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk" Nov 22 05:00:06 crc kubenswrapper[4818]: I1122 05:00:06.194720 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8"] Nov 22 05:00:06 crc kubenswrapper[4818]: I1122 05:00:06.252588 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk"] Nov 22 05:00:06 crc kubenswrapper[4818]: I1122 05:00:06.296816 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="353ed841-0d3b-46df-beec-1906f6d801fd" path="/var/lib/kubelet/pods/353ed841-0d3b-46df-beec-1906f6d801fd/volumes" Nov 22 05:00:06 crc kubenswrapper[4818]: I1122 05:00:06.297702 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5484670-b8e2-40b9-a36b-3fb47bb89421" path="/var/lib/kubelet/pods/d5484670-b8e2-40b9-a36b-3fb47bb89421/volumes" Nov 22 05:00:06 crc kubenswrapper[4818]: I1122 05:00:06.517343 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" event={"ID":"adeee5af-b824-4034-acf1-8cd7a4931b5a","Type":"ContainerStarted","Data":"fe850039fa52e34e9013f3e8c1d5ecaec733f7b4ba7d2465a9628242c01f657d"} Nov 22 05:00:06 crc kubenswrapper[4818]: I1122 05:00:06.518596 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk" event={"ID":"2e7c63aa-6e42-4f75-a595-69df5005d036","Type":"ContainerStarted","Data":"b529e15556c5f12eee7dff0d0a7e316c404564b20584bdad530442b0d792df65"} Nov 22 05:00:07 crc kubenswrapper[4818]: I1122 05:00:07.525646 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" event={"ID":"adeee5af-b824-4034-acf1-8cd7a4931b5a","Type":"ContainerStarted","Data":"59e9720a53ab3a062987bb96ffa4a6b0d1e916acd07528f7133ba22478e51d8e"} Nov 22 05:00:07 crc kubenswrapper[4818]: I1122 05:00:07.525996 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" Nov 22 05:00:07 crc kubenswrapper[4818]: I1122 05:00:07.528147 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk" event={"ID":"2e7c63aa-6e42-4f75-a595-69df5005d036","Type":"ContainerStarted","Data":"6d735c87c96d9eca9a701197157f49ae608b6f89aef0ea34c14470daf7eb1cdd"} Nov 22 05:00:07 crc kubenswrapper[4818]: I1122 05:00:07.528426 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk" Nov 22 05:00:07 crc kubenswrapper[4818]: I1122 05:00:07.535148 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk" Nov 22 05:00:07 crc kubenswrapper[4818]: I1122 05:00:07.535237 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" Nov 22 05:00:07 crc kubenswrapper[4818]: I1122 05:00:07.569895 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7cb6cd57bb-rz7c8" podStartSLOduration=2.56987782 podStartE2EDuration="2.56987782s" podCreationTimestamp="2025-11-22 05:00:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:00:07.567312312 +0000 UTC m=+760.141728859" watchObservedRunningTime="2025-11-22 05:00:07.56987782 +0000 UTC m=+760.144294357" Nov 22 05:00:07 crc kubenswrapper[4818]: I1122 05:00:07.626854 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-774dd8f765-4l6hk" podStartSLOduration=2.626835438 podStartE2EDuration="2.626835438s" podCreationTimestamp="2025-11-22 05:00:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:00:07.625328128 +0000 UTC m=+760.199744655" watchObservedRunningTime="2025-11-22 05:00:07.626835438 +0000 UTC m=+760.201251965" Nov 22 05:00:10 crc kubenswrapper[4818]: I1122 05:00:10.684134 4818 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 22 05:00:21 crc kubenswrapper[4818]: I1122 05:00:21.264540 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:00:21 crc kubenswrapper[4818]: I1122 05:00:21.265207 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:00:51 crc kubenswrapper[4818]: I1122 05:00:51.265129 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:00:51 crc kubenswrapper[4818]: I1122 05:00:51.265979 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:01:21 crc kubenswrapper[4818]: I1122 05:01:21.265075 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:01:21 crc kubenswrapper[4818]: I1122 05:01:21.265879 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:01:21 crc kubenswrapper[4818]: I1122 05:01:21.265947 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 05:01:21 crc kubenswrapper[4818]: I1122 05:01:21.266832 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0c9825f8f0cd7e721d11fe80d62d92bcabce6091d17bb472097983d66a29cf6a"} pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 05:01:21 crc kubenswrapper[4818]: I1122 05:01:21.266934 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" containerID="cri-o://0c9825f8f0cd7e721d11fe80d62d92bcabce6091d17bb472097983d66a29cf6a" gracePeriod=600 Nov 22 05:01:22 crc kubenswrapper[4818]: I1122 05:01:22.011815 4818 generic.go:334] "Generic (PLEG): container finished" podID="0971c9ec-29ce-44eb-9875-ad4664143305" containerID="0c9825f8f0cd7e721d11fe80d62d92bcabce6091d17bb472097983d66a29cf6a" exitCode=0 Nov 22 05:01:22 crc kubenswrapper[4818]: I1122 05:01:22.011877 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerDied","Data":"0c9825f8f0cd7e721d11fe80d62d92bcabce6091d17bb472097983d66a29cf6a"} Nov 22 05:01:22 crc kubenswrapper[4818]: I1122 05:01:22.011938 4818 scope.go:117] "RemoveContainer" containerID="86d6c8c4185dff63c83f0e08d741aef33efa784f30aa076579d91d5ab82bea2b" Nov 22 05:01:23 crc kubenswrapper[4818]: I1122 05:01:23.024087 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"17a029baa87e074670137f3b22e9f934cbfb0d28077e00586b8813874c289d90"} Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.561992 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-248l2"] Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.563319 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-248l2" Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.566076 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.566175 4818 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-tplnv" Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.572467 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-9l45k"] Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.572932 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.573038 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-9l45k" Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.577485 4818 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-xtp6c" Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.583886 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-248l2"] Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.587717 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-9l45k"] Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.607604 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-4bmxf"] Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.608201 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-4bmxf" Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.610040 4818 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-kck5b" Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.623205 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-4bmxf"] Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.709903 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58v82\" (UniqueName: \"kubernetes.io/projected/c62c123f-00fc-4a5e-9649-87121c5f793a-kube-api-access-58v82\") pod \"cert-manager-5b446d88c5-9l45k\" (UID: \"c62c123f-00fc-4a5e-9649-87121c5f793a\") " pod="cert-manager/cert-manager-5b446d88c5-9l45k" Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.709978 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nrfr\" (UniqueName: \"kubernetes.io/projected/77efc3b2-d4e5-426e-ada4-de8c900575f9-kube-api-access-5nrfr\") pod \"cert-manager-webhook-5655c58dd6-4bmxf\" (UID: \"77efc3b2-d4e5-426e-ada4-de8c900575f9\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-4bmxf" Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.710010 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-649dc\" (UniqueName: \"kubernetes.io/projected/de576ad7-b4e6-4047-92d1-dc19f477be33-kube-api-access-649dc\") pod \"cert-manager-cainjector-7f985d654d-248l2\" (UID: \"de576ad7-b4e6-4047-92d1-dc19f477be33\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-248l2" Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.811006 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58v82\" (UniqueName: \"kubernetes.io/projected/c62c123f-00fc-4a5e-9649-87121c5f793a-kube-api-access-58v82\") pod \"cert-manager-5b446d88c5-9l45k\" (UID: \"c62c123f-00fc-4a5e-9649-87121c5f793a\") " pod="cert-manager/cert-manager-5b446d88c5-9l45k" Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.811119 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nrfr\" (UniqueName: \"kubernetes.io/projected/77efc3b2-d4e5-426e-ada4-de8c900575f9-kube-api-access-5nrfr\") pod \"cert-manager-webhook-5655c58dd6-4bmxf\" (UID: \"77efc3b2-d4e5-426e-ada4-de8c900575f9\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-4bmxf" Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.811166 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-649dc\" (UniqueName: \"kubernetes.io/projected/de576ad7-b4e6-4047-92d1-dc19f477be33-kube-api-access-649dc\") pod \"cert-manager-cainjector-7f985d654d-248l2\" (UID: \"de576ad7-b4e6-4047-92d1-dc19f477be33\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-248l2" Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.831716 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58v82\" (UniqueName: \"kubernetes.io/projected/c62c123f-00fc-4a5e-9649-87121c5f793a-kube-api-access-58v82\") pod \"cert-manager-5b446d88c5-9l45k\" (UID: \"c62c123f-00fc-4a5e-9649-87121c5f793a\") " pod="cert-manager/cert-manager-5b446d88c5-9l45k" Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.831819 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-649dc\" (UniqueName: \"kubernetes.io/projected/de576ad7-b4e6-4047-92d1-dc19f477be33-kube-api-access-649dc\") pod \"cert-manager-cainjector-7f985d654d-248l2\" (UID: \"de576ad7-b4e6-4047-92d1-dc19f477be33\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-248l2" Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.832534 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nrfr\" (UniqueName: \"kubernetes.io/projected/77efc3b2-d4e5-426e-ada4-de8c900575f9-kube-api-access-5nrfr\") pod \"cert-manager-webhook-5655c58dd6-4bmxf\" (UID: \"77efc3b2-d4e5-426e-ada4-de8c900575f9\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-4bmxf" Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.890091 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-248l2" Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.903273 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-9l45k" Nov 22 05:02:18 crc kubenswrapper[4818]: I1122 05:02:18.922703 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-4bmxf" Nov 22 05:02:19 crc kubenswrapper[4818]: I1122 05:02:19.180894 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-248l2"] Nov 22 05:02:19 crc kubenswrapper[4818]: I1122 05:02:19.187566 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 05:02:19 crc kubenswrapper[4818]: I1122 05:02:19.211161 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-4bmxf"] Nov 22 05:02:19 crc kubenswrapper[4818]: W1122 05:02:19.214999 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77efc3b2_d4e5_426e_ada4_de8c900575f9.slice/crio-c208e58b55e405cfbd0dd4c6d8f4ca633625cc87ea32e4f12799e16fcb32ef48 WatchSource:0}: Error finding container c208e58b55e405cfbd0dd4c6d8f4ca633625cc87ea32e4f12799e16fcb32ef48: Status 404 returned error can't find the container with id c208e58b55e405cfbd0dd4c6d8f4ca633625cc87ea32e4f12799e16fcb32ef48 Nov 22 05:02:19 crc kubenswrapper[4818]: I1122 05:02:19.341648 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-9l45k"] Nov 22 05:02:19 crc kubenswrapper[4818]: W1122 05:02:19.348861 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc62c123f_00fc_4a5e_9649_87121c5f793a.slice/crio-1252f055a26216e923f711a43f1a33001052239f54c8dca1969552418fb2f7f8 WatchSource:0}: Error finding container 1252f055a26216e923f711a43f1a33001052239f54c8dca1969552418fb2f7f8: Status 404 returned error can't find the container with id 1252f055a26216e923f711a43f1a33001052239f54c8dca1969552418fb2f7f8 Nov 22 05:02:19 crc kubenswrapper[4818]: I1122 05:02:19.380742 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-4bmxf" event={"ID":"77efc3b2-d4e5-426e-ada4-de8c900575f9","Type":"ContainerStarted","Data":"c208e58b55e405cfbd0dd4c6d8f4ca633625cc87ea32e4f12799e16fcb32ef48"} Nov 22 05:02:19 crc kubenswrapper[4818]: I1122 05:02:19.381665 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-9l45k" event={"ID":"c62c123f-00fc-4a5e-9649-87121c5f793a","Type":"ContainerStarted","Data":"1252f055a26216e923f711a43f1a33001052239f54c8dca1969552418fb2f7f8"} Nov 22 05:02:19 crc kubenswrapper[4818]: I1122 05:02:19.382298 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-248l2" event={"ID":"de576ad7-b4e6-4047-92d1-dc19f477be33","Type":"ContainerStarted","Data":"cd029791924485105b11b82645cb4457fac6a0050e42a0fb0b82ffd5d3adcbda"} Nov 22 05:02:26 crc kubenswrapper[4818]: I1122 05:02:26.427767 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-248l2" event={"ID":"de576ad7-b4e6-4047-92d1-dc19f477be33","Type":"ContainerStarted","Data":"9d864fa09664021c7726faec29b4f1c1991d60506a5557877c973cf7de68f957"} Nov 22 05:02:26 crc kubenswrapper[4818]: I1122 05:02:26.430024 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-4bmxf" event={"ID":"77efc3b2-d4e5-426e-ada4-de8c900575f9","Type":"ContainerStarted","Data":"9c1b220912f9d7e35559f06000b818f76e02e15e29d7b7d528a9288a1be0b1c4"} Nov 22 05:02:26 crc kubenswrapper[4818]: I1122 05:02:26.430181 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-4bmxf" Nov 22 05:02:26 crc kubenswrapper[4818]: I1122 05:02:26.431480 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-9l45k" event={"ID":"c62c123f-00fc-4a5e-9649-87121c5f793a","Type":"ContainerStarted","Data":"f81fc1513f5c426b90a82cbd75014e6d5933dd20688fabcc03a87b1e7bbde5a7"} Nov 22 05:02:26 crc kubenswrapper[4818]: I1122 05:02:26.444771 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-248l2" podStartSLOduration=1.522540709 podStartE2EDuration="8.444750871s" podCreationTimestamp="2025-11-22 05:02:18 +0000 UTC" firstStartedPulling="2025-11-22 05:02:19.187337592 +0000 UTC m=+891.761754119" lastFinishedPulling="2025-11-22 05:02:26.109547744 +0000 UTC m=+898.683964281" observedRunningTime="2025-11-22 05:02:26.443974561 +0000 UTC m=+899.018391098" watchObservedRunningTime="2025-11-22 05:02:26.444750871 +0000 UTC m=+899.019167418" Nov 22 05:02:26 crc kubenswrapper[4818]: I1122 05:02:26.461209 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-4bmxf" podStartSLOduration=1.569870194 podStartE2EDuration="8.461180713s" podCreationTimestamp="2025-11-22 05:02:18 +0000 UTC" firstStartedPulling="2025-11-22 05:02:19.218179354 +0000 UTC m=+891.792595881" lastFinishedPulling="2025-11-22 05:02:26.109489863 +0000 UTC m=+898.683906400" observedRunningTime="2025-11-22 05:02:26.458479883 +0000 UTC m=+899.032896420" watchObservedRunningTime="2025-11-22 05:02:26.461180713 +0000 UTC m=+899.035597250" Nov 22 05:02:26 crc kubenswrapper[4818]: I1122 05:02:26.474406 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-9l45k" podStartSLOduration=1.637243984 podStartE2EDuration="8.474382462s" podCreationTimestamp="2025-11-22 05:02:18 +0000 UTC" firstStartedPulling="2025-11-22 05:02:19.351308682 +0000 UTC m=+891.925725199" lastFinishedPulling="2025-11-22 05:02:26.18844713 +0000 UTC m=+898.762863677" observedRunningTime="2025-11-22 05:02:26.472596695 +0000 UTC m=+899.047013232" watchObservedRunningTime="2025-11-22 05:02:26.474382462 +0000 UTC m=+899.048798989" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.158114 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-p8s7r"] Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.159076 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovn-controller" containerID="cri-o://d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc" gracePeriod=30 Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.159105 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="nbdb" containerID="cri-o://8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e" gracePeriod=30 Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.159228 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="northd" containerID="cri-o://7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761" gracePeriod=30 Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.159328 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b" gracePeriod=30 Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.159369 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="sbdb" containerID="cri-o://08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00" gracePeriod=30 Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.159392 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="kube-rbac-proxy-node" containerID="cri-o://f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7" gracePeriod=30 Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.159450 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovn-acl-logging" containerID="cri-o://c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397" gracePeriod=30 Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.197398 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovnkube-controller" containerID="cri-o://66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d" gracePeriod=30 Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.447711 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p8s7r_f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/ovnkube-controller/3.log" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.449627 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p8s7r_f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/ovnkube-controller/3.log" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.449640 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p8s7r_f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/ovn-acl-logging/0.log" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.450105 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p8s7r_f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/ovn-controller/0.log" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.450483 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.451320 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p8s7r_f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/ovn-acl-logging/0.log" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.451746 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p8s7r_f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/ovn-controller/0.log" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.452037 4818 generic.go:334] "Generic (PLEG): container finished" podID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerID="66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d" exitCode=0 Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.452061 4818 generic.go:334] "Generic (PLEG): container finished" podID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerID="08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00" exitCode=0 Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.452070 4818 generic.go:334] "Generic (PLEG): container finished" podID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerID="8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e" exitCode=0 Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.452077 4818 generic.go:334] "Generic (PLEG): container finished" podID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerID="7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761" exitCode=0 Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.452083 4818 generic.go:334] "Generic (PLEG): container finished" podID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerID="051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b" exitCode=0 Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.452090 4818 generic.go:334] "Generic (PLEG): container finished" podID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerID="f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7" exitCode=0 Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.452098 4818 generic.go:334] "Generic (PLEG): container finished" podID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerID="c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397" exitCode=143 Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.452106 4818 generic.go:334] "Generic (PLEG): container finished" podID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerID="d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc" exitCode=143 Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.452106 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerDied","Data":"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.452145 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerDied","Data":"08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.452162 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerDied","Data":"8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.452176 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerDied","Data":"7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.452188 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerDied","Data":"051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.452203 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerDied","Data":"f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.452217 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.452229 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.452237 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.452243 4818 scope.go:117] "RemoveContainer" containerID="66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.452245 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.452995 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453006 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453115 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453127 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453134 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453147 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerDied","Data":"c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453162 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453203 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453210 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453218 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453225 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453232 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453239 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453276 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453286 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453294 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453304 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerDied","Data":"d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453318 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453327 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453362 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453371 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453378 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453386 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453393 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453400 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453409 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453416 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453454 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" event={"ID":"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025","Type":"ContainerDied","Data":"0f7118fd73ae73c6226c848498ad1acd4cf0c3b128a39b62ab344e4193d7973c"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453467 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453476 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453483 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453491 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453498 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453533 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453541 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453548 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453556 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.453563 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.454753 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x865d_d96879f8-2766-4f5e-bc3e-bbf8e5394a94/kube-multus/2.log" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.455274 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x865d_d96879f8-2766-4f5e-bc3e-bbf8e5394a94/kube-multus/1.log" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.455309 4818 generic.go:334] "Generic (PLEG): container finished" podID="d96879f8-2766-4f5e-bc3e-bbf8e5394a94" containerID="6b22db5b4662967112f000f6af50106e3510dc0314e236e402abb833ed2bad54" exitCode=2 Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.455381 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x865d" event={"ID":"d96879f8-2766-4f5e-bc3e-bbf8e5394a94","Type":"ContainerDied","Data":"6b22db5b4662967112f000f6af50106e3510dc0314e236e402abb833ed2bad54"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.455407 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52d352c1148218337557f43c9772c5cc9866398ef20d8b5eb91eeff53c4cecbd"} Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.455667 4818 scope.go:117] "RemoveContainer" containerID="6b22db5b4662967112f000f6af50106e3510dc0314e236e402abb833ed2bad54" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.478978 4818 scope.go:117] "RemoveContainer" containerID="37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.521214 4818 scope.go:117] "RemoveContainer" containerID="08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.525729 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-crtgr"] Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.526462 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovnkube-controller" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.526482 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovnkube-controller" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.526495 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovnkube-controller" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.526504 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovnkube-controller" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.526522 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovn-acl-logging" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.526530 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovn-acl-logging" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.526540 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="kube-rbac-proxy-ovn-metrics" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.526549 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="kube-rbac-proxy-ovn-metrics" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.526740 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovnkube-controller" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.526750 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovnkube-controller" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.526767 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovnkube-controller" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.526776 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovnkube-controller" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.526786 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="northd" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.526795 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="northd" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.526814 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovn-controller" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.526823 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovn-controller" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.526839 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="kube-rbac-proxy-node" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.526847 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="kube-rbac-proxy-node" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.526856 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="sbdb" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.526864 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="sbdb" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.526880 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="nbdb" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.526887 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="nbdb" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.526904 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="kubecfg-setup" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.526912 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="kubecfg-setup" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.527312 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovnkube-controller" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.527327 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="nbdb" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.527338 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="sbdb" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.527361 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovn-acl-logging" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.527376 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovnkube-controller" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.527392 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovnkube-controller" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.527807 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovn-controller" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.527832 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovnkube-controller" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.527848 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="northd" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.527865 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="kube-rbac-proxy-node" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.527874 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="kube-rbac-proxy-ovn-metrics" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.528248 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovnkube-controller" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.528457 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovnkube-controller" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.529129 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" containerName="ovnkube-controller" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.535005 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555294 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-ovnkube-config\") pod \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555330 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-run-openvswitch\") pod \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555352 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-kubelet\") pod \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555366 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-var-lib-cni-networks-ovn-kubernetes\") pod \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555385 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-run-ovn\") pod \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555406 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97n69\" (UniqueName: \"kubernetes.io/projected/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-kube-api-access-97n69\") pod \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555419 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-systemd-units\") pod \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555458 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-var-lib-openvswitch\") pod \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555489 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-ovn-node-metrics-cert\") pod \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555504 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-env-overrides\") pod \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555524 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-slash\") pod \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555542 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-ovnkube-script-lib\") pod \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555539 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" (UID: "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555560 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-log-socket\") pod \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555575 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-node-log\") pod \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555584 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" (UID: "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555588 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-run-ovn-kubernetes\") pod \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555609 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" (UID: "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555627 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-run-netns\") pod \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555668 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-cni-bin\") pod \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555692 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-cni-netd\") pod \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555712 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-run-systemd\") pod \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555739 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-etc-openvswitch\") pod \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\" (UID: \"f8f88a0c-c404-4e82-88c5-3e0f4f2b9025\") " Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.556010 4818 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.556026 4818 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.556037 4818 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.555632 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" (UID: "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.556065 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" (UID: "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.556360 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" (UID: "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.556389 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" (UID: "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.556412 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" (UID: "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.557038 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" (UID: "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.557603 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" (UID: "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.557643 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-slash" (OuterVolumeSpecName: "host-slash") pod "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" (UID: "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.557689 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" (UID: "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.557721 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" (UID: "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.557878 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" (UID: "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.557879 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-log-socket" (OuterVolumeSpecName: "log-socket") pod "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" (UID: "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.557918 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-node-log" (OuterVolumeSpecName: "node-log") pod "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" (UID: "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.557934 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" (UID: "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.561088 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" (UID: "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.561505 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-kube-api-access-97n69" (OuterVolumeSpecName: "kube-api-access-97n69") pod "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" (UID: "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025"). InnerVolumeSpecName "kube-api-access-97n69". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.562580 4818 scope.go:117] "RemoveContainer" containerID="8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.577183 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" (UID: "f8f88a0c-c404-4e82-88c5-3e0f4f2b9025"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.581278 4818 scope.go:117] "RemoveContainer" containerID="7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.597395 4818 scope.go:117] "RemoveContainer" containerID="051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.608908 4818 scope.go:117] "RemoveContainer" containerID="f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.622154 4818 scope.go:117] "RemoveContainer" containerID="c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.633035 4818 scope.go:117] "RemoveContainer" containerID="d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.644291 4818 scope.go:117] "RemoveContainer" containerID="6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.656646 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-node-log\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.656683 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-log-socket\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.656777 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-cni-netd\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.656798 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-cni-bin\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.656834 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-ovnkube-script-lib\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.656861 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-ovnkube-config\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.656933 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.656961 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw9b4\" (UniqueName: \"kubernetes.io/projected/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-kube-api-access-pw9b4\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657010 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-env-overrides\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657044 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-run-ovn\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657075 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-systemd-units\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657098 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-run-netns\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657118 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-run-openvswitch\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657141 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-slash\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657169 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-run-ovn-kubernetes\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657208 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-run-systemd\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657245 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-kubelet\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657279 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-var-lib-openvswitch\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657309 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-ovn-node-metrics-cert\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657331 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-etc-openvswitch\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657399 4818 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657419 4818 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657433 4818 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-slash\") on node \"crc\" DevicePath \"\"" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657445 4818 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657457 4818 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-log-socket\") on node \"crc\" DevicePath \"\"" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657468 4818 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-node-log\") on node \"crc\" DevicePath \"\"" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657479 4818 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657490 4818 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657502 4818 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657513 4818 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657524 4818 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657535 4818 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657545 4818 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657558 4818 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657564 4818 scope.go:117] "RemoveContainer" containerID="66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657572 4818 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657584 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97n69\" (UniqueName: \"kubernetes.io/projected/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-kube-api-access-97n69\") on node \"crc\" DevicePath \"\"" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657597 4818 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.657886 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d\": container with ID starting with 66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d not found: ID does not exist" containerID="66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657907 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d"} err="failed to get container status \"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d\": rpc error: code = NotFound desc = could not find container \"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d\": container with ID starting with 66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.657925 4818 scope.go:117] "RemoveContainer" containerID="37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.658132 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba\": container with ID starting with 37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba not found: ID does not exist" containerID="37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.658150 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba"} err="failed to get container status \"37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba\": rpc error: code = NotFound desc = could not find container \"37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba\": container with ID starting with 37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.658160 4818 scope.go:117] "RemoveContainer" containerID="08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.658382 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\": container with ID starting with 08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00 not found: ID does not exist" containerID="08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.658401 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00"} err="failed to get container status \"08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\": rpc error: code = NotFound desc = could not find container \"08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\": container with ID starting with 08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00 not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.658416 4818 scope.go:117] "RemoveContainer" containerID="8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.658588 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\": container with ID starting with 8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e not found: ID does not exist" containerID="8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.658605 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e"} err="failed to get container status \"8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\": rpc error: code = NotFound desc = could not find container \"8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\": container with ID starting with 8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.658616 4818 scope.go:117] "RemoveContainer" containerID="7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.659029 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\": container with ID starting with 7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761 not found: ID does not exist" containerID="7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.659047 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761"} err="failed to get container status \"7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\": rpc error: code = NotFound desc = could not find container \"7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\": container with ID starting with 7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761 not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.659058 4818 scope.go:117] "RemoveContainer" containerID="051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.659231 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\": container with ID starting with 051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b not found: ID does not exist" containerID="051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.659245 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b"} err="failed to get container status \"051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\": rpc error: code = NotFound desc = could not find container \"051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\": container with ID starting with 051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.659325 4818 scope.go:117] "RemoveContainer" containerID="f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.659591 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\": container with ID starting with f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7 not found: ID does not exist" containerID="f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.659613 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7"} err="failed to get container status \"f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\": rpc error: code = NotFound desc = could not find container \"f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\": container with ID starting with f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7 not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.659630 4818 scope.go:117] "RemoveContainer" containerID="c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.660136 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\": container with ID starting with c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397 not found: ID does not exist" containerID="c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.660175 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397"} err="failed to get container status \"c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\": rpc error: code = NotFound desc = could not find container \"c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\": container with ID starting with c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397 not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.660493 4818 scope.go:117] "RemoveContainer" containerID="d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.661315 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\": container with ID starting with d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc not found: ID does not exist" containerID="d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.661390 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc"} err="failed to get container status \"d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\": rpc error: code = NotFound desc = could not find container \"d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\": container with ID starting with d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.661447 4818 scope.go:117] "RemoveContainer" containerID="6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb" Nov 22 05:02:29 crc kubenswrapper[4818]: E1122 05:02:29.665011 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\": container with ID starting with 6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb not found: ID does not exist" containerID="6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.665045 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb"} err="failed to get container status \"6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\": rpc error: code = NotFound desc = could not find container \"6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\": container with ID starting with 6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.665070 4818 scope.go:117] "RemoveContainer" containerID="66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.665328 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d"} err="failed to get container status \"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d\": rpc error: code = NotFound desc = could not find container \"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d\": container with ID starting with 66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.665344 4818 scope.go:117] "RemoveContainer" containerID="37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.665556 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba"} err="failed to get container status \"37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba\": rpc error: code = NotFound desc = could not find container \"37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba\": container with ID starting with 37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.665574 4818 scope.go:117] "RemoveContainer" containerID="08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.665776 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00"} err="failed to get container status \"08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\": rpc error: code = NotFound desc = could not find container \"08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\": container with ID starting with 08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00 not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.665829 4818 scope.go:117] "RemoveContainer" containerID="8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.666125 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e"} err="failed to get container status \"8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\": rpc error: code = NotFound desc = could not find container \"8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\": container with ID starting with 8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.666149 4818 scope.go:117] "RemoveContainer" containerID="7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.666394 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761"} err="failed to get container status \"7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\": rpc error: code = NotFound desc = could not find container \"7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\": container with ID starting with 7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761 not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.666421 4818 scope.go:117] "RemoveContainer" containerID="051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.666615 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b"} err="failed to get container status \"051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\": rpc error: code = NotFound desc = could not find container \"051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\": container with ID starting with 051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.666636 4818 scope.go:117] "RemoveContainer" containerID="f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.667042 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7"} err="failed to get container status \"f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\": rpc error: code = NotFound desc = could not find container \"f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\": container with ID starting with f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7 not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.667060 4818 scope.go:117] "RemoveContainer" containerID="c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.668209 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397"} err="failed to get container status \"c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\": rpc error: code = NotFound desc = could not find container \"c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\": container with ID starting with c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397 not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.668229 4818 scope.go:117] "RemoveContainer" containerID="d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.677506 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc"} err="failed to get container status \"d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\": rpc error: code = NotFound desc = could not find container \"d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\": container with ID starting with d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.677542 4818 scope.go:117] "RemoveContainer" containerID="6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.678684 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb"} err="failed to get container status \"6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\": rpc error: code = NotFound desc = could not find container \"6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\": container with ID starting with 6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.678705 4818 scope.go:117] "RemoveContainer" containerID="66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.678932 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d"} err="failed to get container status \"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d\": rpc error: code = NotFound desc = could not find container \"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d\": container with ID starting with 66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.678953 4818 scope.go:117] "RemoveContainer" containerID="37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.679199 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba"} err="failed to get container status \"37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba\": rpc error: code = NotFound desc = could not find container \"37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba\": container with ID starting with 37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.679235 4818 scope.go:117] "RemoveContainer" containerID="08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.679828 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00"} err="failed to get container status \"08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\": rpc error: code = NotFound desc = could not find container \"08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\": container with ID starting with 08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00 not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.679844 4818 scope.go:117] "RemoveContainer" containerID="8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.683434 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e"} err="failed to get container status \"8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\": rpc error: code = NotFound desc = could not find container \"8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\": container with ID starting with 8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.683484 4818 scope.go:117] "RemoveContainer" containerID="7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.684116 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761"} err="failed to get container status \"7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\": rpc error: code = NotFound desc = could not find container \"7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\": container with ID starting with 7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761 not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.684167 4818 scope.go:117] "RemoveContainer" containerID="051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.685542 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b"} err="failed to get container status \"051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\": rpc error: code = NotFound desc = could not find container \"051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\": container with ID starting with 051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.685595 4818 scope.go:117] "RemoveContainer" containerID="f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.685985 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7"} err="failed to get container status \"f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\": rpc error: code = NotFound desc = could not find container \"f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\": container with ID starting with f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7 not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.686021 4818 scope.go:117] "RemoveContainer" containerID="c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.686319 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397"} err="failed to get container status \"c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\": rpc error: code = NotFound desc = could not find container \"c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\": container with ID starting with c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397 not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.686344 4818 scope.go:117] "RemoveContainer" containerID="d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.686584 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc"} err="failed to get container status \"d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\": rpc error: code = NotFound desc = could not find container \"d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\": container with ID starting with d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.686606 4818 scope.go:117] "RemoveContainer" containerID="6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.686960 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb"} err="failed to get container status \"6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\": rpc error: code = NotFound desc = could not find container \"6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\": container with ID starting with 6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.686979 4818 scope.go:117] "RemoveContainer" containerID="66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.687426 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d"} err="failed to get container status \"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d\": rpc error: code = NotFound desc = could not find container \"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d\": container with ID starting with 66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.687445 4818 scope.go:117] "RemoveContainer" containerID="37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.687722 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba"} err="failed to get container status \"37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba\": rpc error: code = NotFound desc = could not find container \"37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba\": container with ID starting with 37312ec74f1aea14eecd8bedb8925d086b3ac5f64af0e8abec48db61ed3976ba not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.687742 4818 scope.go:117] "RemoveContainer" containerID="08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.688032 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00"} err="failed to get container status \"08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\": rpc error: code = NotFound desc = could not find container \"08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00\": container with ID starting with 08f1bcc7ef77f093bb83206cf217e605c749991f8a8d49ea59c239e015c28a00 not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.688052 4818 scope.go:117] "RemoveContainer" containerID="8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.688289 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e"} err="failed to get container status \"8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\": rpc error: code = NotFound desc = could not find container \"8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e\": container with ID starting with 8edb8755e3d8f91304ae02abcfb55111a9dd9b017c5cb0df8d7857f3f6c28e7e not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.688323 4818 scope.go:117] "RemoveContainer" containerID="7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.688588 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761"} err="failed to get container status \"7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\": rpc error: code = NotFound desc = could not find container \"7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761\": container with ID starting with 7e14ccc7c53957e22f7330b19604accbbc4dad265b05f6e2ad20ea2634ea8761 not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.688608 4818 scope.go:117] "RemoveContainer" containerID="051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.689467 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b"} err="failed to get container status \"051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\": rpc error: code = NotFound desc = could not find container \"051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b\": container with ID starting with 051341c6c0c92ffd91097e54e62b435d5bcb04cb65a61e6dcfaef66ed0bb8b8b not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.689505 4818 scope.go:117] "RemoveContainer" containerID="f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.689900 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7"} err="failed to get container status \"f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\": rpc error: code = NotFound desc = could not find container \"f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7\": container with ID starting with f0a327e3ea6ebef9148516c52a3f11c0e067b2b7edd503f8a3abc3c71c49dbd7 not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.689927 4818 scope.go:117] "RemoveContainer" containerID="c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.691012 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397"} err="failed to get container status \"c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\": rpc error: code = NotFound desc = could not find container \"c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397\": container with ID starting with c8cd16ca389296473741e0beab922a30c74af57276fb7e101f2180d8587ad397 not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.691029 4818 scope.go:117] "RemoveContainer" containerID="d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.692235 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc"} err="failed to get container status \"d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\": rpc error: code = NotFound desc = could not find container \"d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc\": container with ID starting with d3e1f3c512337e1ebf94ce064763495252bcb52978e4ac228ad66ab331455ecc not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.692287 4818 scope.go:117] "RemoveContainer" containerID="6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.692695 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb"} err="failed to get container status \"6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\": rpc error: code = NotFound desc = could not find container \"6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb\": container with ID starting with 6130fe8009a5753ef9c73cdf73573f1be699033298e03a6fb41eb1f23274fffb not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.692719 4818 scope.go:117] "RemoveContainer" containerID="66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.693426 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d"} err="failed to get container status \"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d\": rpc error: code = NotFound desc = could not find container \"66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d\": container with ID starting with 66f4ab05b9e608747f1ef2148dbb230d2d53438b83bb47554153754e40e90f6d not found: ID does not exist" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758495 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-run-ovn-kubernetes\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758540 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-run-systemd\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758570 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-kubelet\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758583 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-var-lib-openvswitch\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758604 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-ovn-node-metrics-cert\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758626 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-etc-openvswitch\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758639 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-node-log\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758654 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-log-socket\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758677 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-cni-netd\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758693 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-cni-bin\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758718 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-ovnkube-script-lib\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758735 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-ovnkube-config\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758769 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758785 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw9b4\" (UniqueName: \"kubernetes.io/projected/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-kube-api-access-pw9b4\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758802 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-env-overrides\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758820 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-run-ovn\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758838 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-systemd-units\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758854 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-run-netns\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758869 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-run-openvswitch\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758885 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-slash\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758956 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-slash\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.758998 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-run-ovn-kubernetes\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.759019 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-run-systemd\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.759039 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-kubelet\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.759061 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-var-lib-openvswitch\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.759919 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-log-socket\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.759961 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-etc-openvswitch\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.759991 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-node-log\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.759999 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-cni-bin\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.760040 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-run-ovn\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.760065 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-run-netns\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.760112 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-run-openvswitch\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.760080 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.760009 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-host-cni-netd\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.760491 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-systemd-units\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.760512 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-ovnkube-config\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.760526 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-env-overrides\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.760771 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-ovnkube-script-lib\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.762809 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-ovn-node-metrics-cert\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.781582 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw9b4\" (UniqueName: \"kubernetes.io/projected/71f2dc21-d4bd-42f7-b9cf-88b895c2bece-kube-api-access-pw9b4\") pod \"ovnkube-node-crtgr\" (UID: \"71f2dc21-d4bd-42f7-b9cf-88b895c2bece\") " pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: I1122 05:02:29.861736 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:29 crc kubenswrapper[4818]: W1122 05:02:29.879181 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71f2dc21_d4bd_42f7_b9cf_88b895c2bece.slice/crio-9298da641254ce361df4abed06deee8aa46809185a7d3fcc93f863afdbc24cfd WatchSource:0}: Error finding container 9298da641254ce361df4abed06deee8aa46809185a7d3fcc93f863afdbc24cfd: Status 404 returned error can't find the container with id 9298da641254ce361df4abed06deee8aa46809185a7d3fcc93f863afdbc24cfd Nov 22 05:02:30 crc kubenswrapper[4818]: I1122 05:02:30.474854 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p8s7r" Nov 22 05:02:30 crc kubenswrapper[4818]: I1122 05:02:30.477561 4818 generic.go:334] "Generic (PLEG): container finished" podID="71f2dc21-d4bd-42f7-b9cf-88b895c2bece" containerID="75d75061d224d5afd565813994cac560240a057095f51590709fce845a09a981" exitCode=0 Nov 22 05:02:30 crc kubenswrapper[4818]: I1122 05:02:30.477647 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" event={"ID":"71f2dc21-d4bd-42f7-b9cf-88b895c2bece","Type":"ContainerDied","Data":"75d75061d224d5afd565813994cac560240a057095f51590709fce845a09a981"} Nov 22 05:02:30 crc kubenswrapper[4818]: I1122 05:02:30.477690 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" event={"ID":"71f2dc21-d4bd-42f7-b9cf-88b895c2bece","Type":"ContainerStarted","Data":"9298da641254ce361df4abed06deee8aa46809185a7d3fcc93f863afdbc24cfd"} Nov 22 05:02:30 crc kubenswrapper[4818]: I1122 05:02:30.483192 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x865d_d96879f8-2766-4f5e-bc3e-bbf8e5394a94/kube-multus/2.log" Nov 22 05:02:30 crc kubenswrapper[4818]: I1122 05:02:30.485477 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x865d_d96879f8-2766-4f5e-bc3e-bbf8e5394a94/kube-multus/1.log" Nov 22 05:02:30 crc kubenswrapper[4818]: I1122 05:02:30.485547 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x865d" event={"ID":"d96879f8-2766-4f5e-bc3e-bbf8e5394a94","Type":"ContainerStarted","Data":"173bbbbaa1cf2ae71609a2332223c5047f44a875ad6876444c27527137d5702e"} Nov 22 05:02:30 crc kubenswrapper[4818]: I1122 05:02:30.534782 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-p8s7r"] Nov 22 05:02:30 crc kubenswrapper[4818]: I1122 05:02:30.535115 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-p8s7r"] Nov 22 05:02:31 crc kubenswrapper[4818]: I1122 05:02:31.499626 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" event={"ID":"71f2dc21-d4bd-42f7-b9cf-88b895c2bece","Type":"ContainerStarted","Data":"f123e636d978a76340d6981af736199a04b937727d3595fb890934686e06c7f9"} Nov 22 05:02:31 crc kubenswrapper[4818]: I1122 05:02:31.500214 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" event={"ID":"71f2dc21-d4bd-42f7-b9cf-88b895c2bece","Type":"ContainerStarted","Data":"fc68e34702202874ba37929f1706225ad5b9abbcefbb63d213903e390ef36a51"} Nov 22 05:02:31 crc kubenswrapper[4818]: I1122 05:02:31.500235 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" event={"ID":"71f2dc21-d4bd-42f7-b9cf-88b895c2bece","Type":"ContainerStarted","Data":"95e61d0784efe2032c423cbf46b846bf93fe91b92c972be2ebbde08b6da40989"} Nov 22 05:02:31 crc kubenswrapper[4818]: I1122 05:02:31.500277 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" event={"ID":"71f2dc21-d4bd-42f7-b9cf-88b895c2bece","Type":"ContainerStarted","Data":"cf7db0c0a180fe970b4d144eff442726709e732779e8eb10c6c4ea1549b9d8ca"} Nov 22 05:02:31 crc kubenswrapper[4818]: I1122 05:02:31.500295 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" event={"ID":"71f2dc21-d4bd-42f7-b9cf-88b895c2bece","Type":"ContainerStarted","Data":"39812e0faf72abfbeeac275e94dafb96554037b950026bdac3fe561242250674"} Nov 22 05:02:31 crc kubenswrapper[4818]: I1122 05:02:31.500310 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" event={"ID":"71f2dc21-d4bd-42f7-b9cf-88b895c2bece","Type":"ContainerStarted","Data":"11a906c69750869b452b7522b7c3691a6258532171d3b6e4664c0c0163d7422a"} Nov 22 05:02:32 crc kubenswrapper[4818]: I1122 05:02:32.305621 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8f88a0c-c404-4e82-88c5-3e0f4f2b9025" path="/var/lib/kubelet/pods/f8f88a0c-c404-4e82-88c5-3e0f4f2b9025/volumes" Nov 22 05:02:33 crc kubenswrapper[4818]: I1122 05:02:33.521756 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" event={"ID":"71f2dc21-d4bd-42f7-b9cf-88b895c2bece","Type":"ContainerStarted","Data":"e28526a403eb646f93613a7627f5bdf217c799c5f0040830d7a4be76eaf8b6a9"} Nov 22 05:02:33 crc kubenswrapper[4818]: I1122 05:02:33.925349 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-4bmxf" Nov 22 05:02:36 crc kubenswrapper[4818]: I1122 05:02:36.550789 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" event={"ID":"71f2dc21-d4bd-42f7-b9cf-88b895c2bece","Type":"ContainerStarted","Data":"9bd0014ad1fa539d1b7068cba604331fdb94df4191ea419d8f6771628c889d0f"} Nov 22 05:02:37 crc kubenswrapper[4818]: I1122 05:02:37.556158 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:37 crc kubenswrapper[4818]: I1122 05:02:37.557055 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:37 crc kubenswrapper[4818]: I1122 05:02:37.557136 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:37 crc kubenswrapper[4818]: I1122 05:02:37.589694 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:37 crc kubenswrapper[4818]: I1122 05:02:37.592996 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:02:37 crc kubenswrapper[4818]: I1122 05:02:37.598130 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" podStartSLOduration=8.598112534 podStartE2EDuration="8.598112534s" podCreationTimestamp="2025-11-22 05:02:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:02:37.595301062 +0000 UTC m=+910.169717589" watchObservedRunningTime="2025-11-22 05:02:37.598112534 +0000 UTC m=+910.172529051" Nov 22 05:02:59 crc kubenswrapper[4818]: I1122 05:02:59.897475 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-crtgr" Nov 22 05:03:22 crc kubenswrapper[4818]: I1122 05:03:22.820052 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd"] Nov 22 05:03:22 crc kubenswrapper[4818]: I1122 05:03:22.821807 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd" Nov 22 05:03:22 crc kubenswrapper[4818]: I1122 05:03:22.824227 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 22 05:03:22 crc kubenswrapper[4818]: I1122 05:03:22.830722 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd"] Nov 22 05:03:22 crc kubenswrapper[4818]: I1122 05:03:22.998000 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cc4bf9a2-88ba-4882-9908-f1e2f27c82ac-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd\" (UID: \"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd" Nov 22 05:03:22 crc kubenswrapper[4818]: I1122 05:03:22.998536 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp9rs\" (UniqueName: \"kubernetes.io/projected/cc4bf9a2-88ba-4882-9908-f1e2f27c82ac-kube-api-access-jp9rs\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd\" (UID: \"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd" Nov 22 05:03:22 crc kubenswrapper[4818]: I1122 05:03:22.998563 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cc4bf9a2-88ba-4882-9908-f1e2f27c82ac-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd\" (UID: \"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd" Nov 22 05:03:23 crc kubenswrapper[4818]: I1122 05:03:23.099225 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp9rs\" (UniqueName: \"kubernetes.io/projected/cc4bf9a2-88ba-4882-9908-f1e2f27c82ac-kube-api-access-jp9rs\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd\" (UID: \"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd" Nov 22 05:03:23 crc kubenswrapper[4818]: I1122 05:03:23.099518 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cc4bf9a2-88ba-4882-9908-f1e2f27c82ac-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd\" (UID: \"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd" Nov 22 05:03:23 crc kubenswrapper[4818]: I1122 05:03:23.099667 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cc4bf9a2-88ba-4882-9908-f1e2f27c82ac-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd\" (UID: \"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd" Nov 22 05:03:23 crc kubenswrapper[4818]: I1122 05:03:23.099991 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cc4bf9a2-88ba-4882-9908-f1e2f27c82ac-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd\" (UID: \"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd" Nov 22 05:03:23 crc kubenswrapper[4818]: I1122 05:03:23.100043 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cc4bf9a2-88ba-4882-9908-f1e2f27c82ac-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd\" (UID: \"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd" Nov 22 05:03:23 crc kubenswrapper[4818]: I1122 05:03:23.123617 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp9rs\" (UniqueName: \"kubernetes.io/projected/cc4bf9a2-88ba-4882-9908-f1e2f27c82ac-kube-api-access-jp9rs\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd\" (UID: \"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd" Nov 22 05:03:23 crc kubenswrapper[4818]: I1122 05:03:23.145890 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd" Nov 22 05:03:23 crc kubenswrapper[4818]: I1122 05:03:23.376009 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd"] Nov 22 05:03:23 crc kubenswrapper[4818]: I1122 05:03:23.959760 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd" event={"ID":"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac","Type":"ContainerStarted","Data":"3375ee04359cd14664640be4164f721fa59821556a6e0075cd48db9683bbd185"} Nov 22 05:03:23 crc kubenswrapper[4818]: I1122 05:03:23.959801 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd" event={"ID":"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac","Type":"ContainerStarted","Data":"eea755a102b33ca124747b4f3884c6452000098fa3589a7b6e4525ae19980fb1"} Nov 22 05:03:24 crc kubenswrapper[4818]: I1122 05:03:24.436924 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-trnm7"] Nov 22 05:03:24 crc kubenswrapper[4818]: I1122 05:03:24.452295 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-trnm7" Nov 22 05:03:24 crc kubenswrapper[4818]: I1122 05:03:24.463912 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-trnm7"] Nov 22 05:03:24 crc kubenswrapper[4818]: I1122 05:03:24.515467 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40cb0642-0b94-4089-bf50-f01484850cb5-catalog-content\") pod \"redhat-operators-trnm7\" (UID: \"40cb0642-0b94-4089-bf50-f01484850cb5\") " pod="openshift-marketplace/redhat-operators-trnm7" Nov 22 05:03:24 crc kubenswrapper[4818]: I1122 05:03:24.515623 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6td5\" (UniqueName: \"kubernetes.io/projected/40cb0642-0b94-4089-bf50-f01484850cb5-kube-api-access-p6td5\") pod \"redhat-operators-trnm7\" (UID: \"40cb0642-0b94-4089-bf50-f01484850cb5\") " pod="openshift-marketplace/redhat-operators-trnm7" Nov 22 05:03:24 crc kubenswrapper[4818]: I1122 05:03:24.515664 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40cb0642-0b94-4089-bf50-f01484850cb5-utilities\") pod \"redhat-operators-trnm7\" (UID: \"40cb0642-0b94-4089-bf50-f01484850cb5\") " pod="openshift-marketplace/redhat-operators-trnm7" Nov 22 05:03:24 crc kubenswrapper[4818]: I1122 05:03:24.616363 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6td5\" (UniqueName: \"kubernetes.io/projected/40cb0642-0b94-4089-bf50-f01484850cb5-kube-api-access-p6td5\") pod \"redhat-operators-trnm7\" (UID: \"40cb0642-0b94-4089-bf50-f01484850cb5\") " pod="openshift-marketplace/redhat-operators-trnm7" Nov 22 05:03:24 crc kubenswrapper[4818]: I1122 05:03:24.616412 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40cb0642-0b94-4089-bf50-f01484850cb5-utilities\") pod \"redhat-operators-trnm7\" (UID: \"40cb0642-0b94-4089-bf50-f01484850cb5\") " pod="openshift-marketplace/redhat-operators-trnm7" Nov 22 05:03:24 crc kubenswrapper[4818]: I1122 05:03:24.616440 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40cb0642-0b94-4089-bf50-f01484850cb5-catalog-content\") pod \"redhat-operators-trnm7\" (UID: \"40cb0642-0b94-4089-bf50-f01484850cb5\") " pod="openshift-marketplace/redhat-operators-trnm7" Nov 22 05:03:24 crc kubenswrapper[4818]: I1122 05:03:24.616848 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40cb0642-0b94-4089-bf50-f01484850cb5-catalog-content\") pod \"redhat-operators-trnm7\" (UID: \"40cb0642-0b94-4089-bf50-f01484850cb5\") " pod="openshift-marketplace/redhat-operators-trnm7" Nov 22 05:03:24 crc kubenswrapper[4818]: I1122 05:03:24.617367 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40cb0642-0b94-4089-bf50-f01484850cb5-utilities\") pod \"redhat-operators-trnm7\" (UID: \"40cb0642-0b94-4089-bf50-f01484850cb5\") " pod="openshift-marketplace/redhat-operators-trnm7" Nov 22 05:03:24 crc kubenswrapper[4818]: I1122 05:03:24.649162 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6td5\" (UniqueName: \"kubernetes.io/projected/40cb0642-0b94-4089-bf50-f01484850cb5-kube-api-access-p6td5\") pod \"redhat-operators-trnm7\" (UID: \"40cb0642-0b94-4089-bf50-f01484850cb5\") " pod="openshift-marketplace/redhat-operators-trnm7" Nov 22 05:03:24 crc kubenswrapper[4818]: I1122 05:03:24.773295 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-trnm7" Nov 22 05:03:24 crc kubenswrapper[4818]: I1122 05:03:24.968306 4818 generic.go:334] "Generic (PLEG): container finished" podID="cc4bf9a2-88ba-4882-9908-f1e2f27c82ac" containerID="3375ee04359cd14664640be4164f721fa59821556a6e0075cd48db9683bbd185" exitCode=0 Nov 22 05:03:24 crc kubenswrapper[4818]: I1122 05:03:24.968343 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd" event={"ID":"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac","Type":"ContainerDied","Data":"3375ee04359cd14664640be4164f721fa59821556a6e0075cd48db9683bbd185"} Nov 22 05:03:25 crc kubenswrapper[4818]: I1122 05:03:25.023857 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-trnm7"] Nov 22 05:03:25 crc kubenswrapper[4818]: W1122 05:03:25.032155 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40cb0642_0b94_4089_bf50_f01484850cb5.slice/crio-cdbd90208b9d64fb76f051512d2520be1049040e45adfc7916c08ee40e8bd312 WatchSource:0}: Error finding container cdbd90208b9d64fb76f051512d2520be1049040e45adfc7916c08ee40e8bd312: Status 404 returned error can't find the container with id cdbd90208b9d64fb76f051512d2520be1049040e45adfc7916c08ee40e8bd312 Nov 22 05:03:25 crc kubenswrapper[4818]: I1122 05:03:25.979523 4818 generic.go:334] "Generic (PLEG): container finished" podID="40cb0642-0b94-4089-bf50-f01484850cb5" containerID="e9d1831e8d42c03e0f00d51874b3b74426511ada5b2e17189672f0d8cafd4f89" exitCode=0 Nov 22 05:03:25 crc kubenswrapper[4818]: I1122 05:03:25.979640 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-trnm7" event={"ID":"40cb0642-0b94-4089-bf50-f01484850cb5","Type":"ContainerDied","Data":"e9d1831e8d42c03e0f00d51874b3b74426511ada5b2e17189672f0d8cafd4f89"} Nov 22 05:03:25 crc kubenswrapper[4818]: I1122 05:03:25.980063 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-trnm7" event={"ID":"40cb0642-0b94-4089-bf50-f01484850cb5","Type":"ContainerStarted","Data":"cdbd90208b9d64fb76f051512d2520be1049040e45adfc7916c08ee40e8bd312"} Nov 22 05:03:28 crc kubenswrapper[4818]: I1122 05:03:28.719388 4818 scope.go:117] "RemoveContainer" containerID="52d352c1148218337557f43c9772c5cc9866398ef20d8b5eb91eeff53c4cecbd" Nov 22 05:03:30 crc kubenswrapper[4818]: I1122 05:03:30.009693 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x865d_d96879f8-2766-4f5e-bc3e-bbf8e5394a94/kube-multus/2.log" Nov 22 05:03:32 crc kubenswrapper[4818]: I1122 05:03:32.023570 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd" event={"ID":"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac","Type":"ContainerDied","Data":"b368f67a9313e2909e4562ed53e1f744955d40f4edcf3a394e6088b622b4bf67"} Nov 22 05:03:32 crc kubenswrapper[4818]: I1122 05:03:32.023730 4818 generic.go:334] "Generic (PLEG): container finished" podID="cc4bf9a2-88ba-4882-9908-f1e2f27c82ac" containerID="b368f67a9313e2909e4562ed53e1f744955d40f4edcf3a394e6088b622b4bf67" exitCode=0 Nov 22 05:03:32 crc kubenswrapper[4818]: I1122 05:03:32.026330 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-trnm7" event={"ID":"40cb0642-0b94-4089-bf50-f01484850cb5","Type":"ContainerStarted","Data":"45cf788a70df09717c4e968821098d6d01fa5de366704686190b23f2586f8d83"} Nov 22 05:03:33 crc kubenswrapper[4818]: I1122 05:03:33.033296 4818 generic.go:334] "Generic (PLEG): container finished" podID="cc4bf9a2-88ba-4882-9908-f1e2f27c82ac" containerID="26b6342460a62971ca79e63d7d3c07d7327724792ab2741306cc18302d882fb6" exitCode=0 Nov 22 05:03:33 crc kubenswrapper[4818]: I1122 05:03:33.033349 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd" event={"ID":"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac","Type":"ContainerDied","Data":"26b6342460a62971ca79e63d7d3c07d7327724792ab2741306cc18302d882fb6"} Nov 22 05:03:33 crc kubenswrapper[4818]: I1122 05:03:33.036338 4818 generic.go:334] "Generic (PLEG): container finished" podID="40cb0642-0b94-4089-bf50-f01484850cb5" containerID="45cf788a70df09717c4e968821098d6d01fa5de366704686190b23f2586f8d83" exitCode=0 Nov 22 05:03:33 crc kubenswrapper[4818]: I1122 05:03:33.036378 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-trnm7" event={"ID":"40cb0642-0b94-4089-bf50-f01484850cb5","Type":"ContainerDied","Data":"45cf788a70df09717c4e968821098d6d01fa5de366704686190b23f2586f8d83"} Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.039781 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-msj6n"] Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.041191 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-msj6n" Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.047092 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-trnm7" event={"ID":"40cb0642-0b94-4089-bf50-f01484850cb5","Type":"ContainerStarted","Data":"6b0e799d2bc2e51a5e49890ae2b39c018723ac913f0f9be6926c9a064ea5581d"} Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.060076 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-msj6n"] Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.147500 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfzj9\" (UniqueName: \"kubernetes.io/projected/a988ba9e-742e-48e3-80de-0b7017216cf7-kube-api-access-hfzj9\") pod \"redhat-marketplace-msj6n\" (UID: \"a988ba9e-742e-48e3-80de-0b7017216cf7\") " pod="openshift-marketplace/redhat-marketplace-msj6n" Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.147894 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a988ba9e-742e-48e3-80de-0b7017216cf7-catalog-content\") pod \"redhat-marketplace-msj6n\" (UID: \"a988ba9e-742e-48e3-80de-0b7017216cf7\") " pod="openshift-marketplace/redhat-marketplace-msj6n" Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.147915 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a988ba9e-742e-48e3-80de-0b7017216cf7-utilities\") pod \"redhat-marketplace-msj6n\" (UID: \"a988ba9e-742e-48e3-80de-0b7017216cf7\") " pod="openshift-marketplace/redhat-marketplace-msj6n" Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.249569 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfzj9\" (UniqueName: \"kubernetes.io/projected/a988ba9e-742e-48e3-80de-0b7017216cf7-kube-api-access-hfzj9\") pod \"redhat-marketplace-msj6n\" (UID: \"a988ba9e-742e-48e3-80de-0b7017216cf7\") " pod="openshift-marketplace/redhat-marketplace-msj6n" Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.249646 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a988ba9e-742e-48e3-80de-0b7017216cf7-catalog-content\") pod \"redhat-marketplace-msj6n\" (UID: \"a988ba9e-742e-48e3-80de-0b7017216cf7\") " pod="openshift-marketplace/redhat-marketplace-msj6n" Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.249677 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a988ba9e-742e-48e3-80de-0b7017216cf7-utilities\") pod \"redhat-marketplace-msj6n\" (UID: \"a988ba9e-742e-48e3-80de-0b7017216cf7\") " pod="openshift-marketplace/redhat-marketplace-msj6n" Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.250428 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a988ba9e-742e-48e3-80de-0b7017216cf7-utilities\") pod \"redhat-marketplace-msj6n\" (UID: \"a988ba9e-742e-48e3-80de-0b7017216cf7\") " pod="openshift-marketplace/redhat-marketplace-msj6n" Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.250452 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a988ba9e-742e-48e3-80de-0b7017216cf7-catalog-content\") pod \"redhat-marketplace-msj6n\" (UID: \"a988ba9e-742e-48e3-80de-0b7017216cf7\") " pod="openshift-marketplace/redhat-marketplace-msj6n" Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.274401 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfzj9\" (UniqueName: \"kubernetes.io/projected/a988ba9e-742e-48e3-80de-0b7017216cf7-kube-api-access-hfzj9\") pod \"redhat-marketplace-msj6n\" (UID: \"a988ba9e-742e-48e3-80de-0b7017216cf7\") " pod="openshift-marketplace/redhat-marketplace-msj6n" Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.316265 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd" Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.335328 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-trnm7" podStartSLOduration=2.801018231 podStartE2EDuration="10.335313197s" podCreationTimestamp="2025-11-22 05:03:24 +0000 UTC" firstStartedPulling="2025-11-22 05:03:25.982932111 +0000 UTC m=+958.557348648" lastFinishedPulling="2025-11-22 05:03:33.517227077 +0000 UTC m=+966.091643614" observedRunningTime="2025-11-22 05:03:34.102688009 +0000 UTC m=+966.677104566" watchObservedRunningTime="2025-11-22 05:03:34.335313197 +0000 UTC m=+966.909729724" Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.362972 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-msj6n" Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.451994 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cc4bf9a2-88ba-4882-9908-f1e2f27c82ac-bundle\") pod \"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac\" (UID: \"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac\") " Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.452046 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jp9rs\" (UniqueName: \"kubernetes.io/projected/cc4bf9a2-88ba-4882-9908-f1e2f27c82ac-kube-api-access-jp9rs\") pod \"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac\" (UID: \"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac\") " Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.452140 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cc4bf9a2-88ba-4882-9908-f1e2f27c82ac-util\") pod \"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac\" (UID: \"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac\") " Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.453882 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc4bf9a2-88ba-4882-9908-f1e2f27c82ac-bundle" (OuterVolumeSpecName: "bundle") pod "cc4bf9a2-88ba-4882-9908-f1e2f27c82ac" (UID: "cc4bf9a2-88ba-4882-9908-f1e2f27c82ac"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.460862 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc4bf9a2-88ba-4882-9908-f1e2f27c82ac-kube-api-access-jp9rs" (OuterVolumeSpecName: "kube-api-access-jp9rs") pod "cc4bf9a2-88ba-4882-9908-f1e2f27c82ac" (UID: "cc4bf9a2-88ba-4882-9908-f1e2f27c82ac"). InnerVolumeSpecName "kube-api-access-jp9rs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.466803 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc4bf9a2-88ba-4882-9908-f1e2f27c82ac-util" (OuterVolumeSpecName: "util") pod "cc4bf9a2-88ba-4882-9908-f1e2f27c82ac" (UID: "cc4bf9a2-88ba-4882-9908-f1e2f27c82ac"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.553042 4818 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cc4bf9a2-88ba-4882-9908-f1e2f27c82ac-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.553072 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jp9rs\" (UniqueName: \"kubernetes.io/projected/cc4bf9a2-88ba-4882-9908-f1e2f27c82ac-kube-api-access-jp9rs\") on node \"crc\" DevicePath \"\"" Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.553081 4818 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cc4bf9a2-88ba-4882-9908-f1e2f27c82ac-util\") on node \"crc\" DevicePath \"\"" Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.556876 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-msj6n"] Nov 22 05:03:34 crc kubenswrapper[4818]: W1122 05:03:34.564688 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda988ba9e_742e_48e3_80de_0b7017216cf7.slice/crio-707ee3b2f285412ea3ef6d0ea6263ae48123c864edc1ff5be3806618e45f7110 WatchSource:0}: Error finding container 707ee3b2f285412ea3ef6d0ea6263ae48123c864edc1ff5be3806618e45f7110: Status 404 returned error can't find the container with id 707ee3b2f285412ea3ef6d0ea6263ae48123c864edc1ff5be3806618e45f7110 Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.774182 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-trnm7" Nov 22 05:03:34 crc kubenswrapper[4818]: I1122 05:03:34.774355 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-trnm7" Nov 22 05:03:35 crc kubenswrapper[4818]: I1122 05:03:35.053274 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msj6n" event={"ID":"a988ba9e-742e-48e3-80de-0b7017216cf7","Type":"ContainerStarted","Data":"bc8a87598ee0971a39bd2a39323760f04170a5c9c908e26233b3c27694a02ea0"} Nov 22 05:03:35 crc kubenswrapper[4818]: I1122 05:03:35.053319 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msj6n" event={"ID":"a988ba9e-742e-48e3-80de-0b7017216cf7","Type":"ContainerStarted","Data":"707ee3b2f285412ea3ef6d0ea6263ae48123c864edc1ff5be3806618e45f7110"} Nov 22 05:03:35 crc kubenswrapper[4818]: I1122 05:03:35.055823 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd" Nov 22 05:03:35 crc kubenswrapper[4818]: I1122 05:03:35.056910 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd" event={"ID":"cc4bf9a2-88ba-4882-9908-f1e2f27c82ac","Type":"ContainerDied","Data":"eea755a102b33ca124747b4f3884c6452000098fa3589a7b6e4525ae19980fb1"} Nov 22 05:03:35 crc kubenswrapper[4818]: I1122 05:03:35.056944 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eea755a102b33ca124747b4f3884c6452000098fa3589a7b6e4525ae19980fb1" Nov 22 05:03:35 crc kubenswrapper[4818]: I1122 05:03:35.827667 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-trnm7" podUID="40cb0642-0b94-4089-bf50-f01484850cb5" containerName="registry-server" probeResult="failure" output=< Nov 22 05:03:35 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Nov 22 05:03:35 crc kubenswrapper[4818]: > Nov 22 05:03:36 crc kubenswrapper[4818]: I1122 05:03:36.061360 4818 generic.go:334] "Generic (PLEG): container finished" podID="a988ba9e-742e-48e3-80de-0b7017216cf7" containerID="bc8a87598ee0971a39bd2a39323760f04170a5c9c908e26233b3c27694a02ea0" exitCode=0 Nov 22 05:03:36 crc kubenswrapper[4818]: I1122 05:03:36.061403 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msj6n" event={"ID":"a988ba9e-742e-48e3-80de-0b7017216cf7","Type":"ContainerDied","Data":"bc8a87598ee0971a39bd2a39323760f04170a5c9c908e26233b3c27694a02ea0"} Nov 22 05:03:38 crc kubenswrapper[4818]: I1122 05:03:38.642019 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-r2wjf"] Nov 22 05:03:38 crc kubenswrapper[4818]: E1122 05:03:38.643603 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc4bf9a2-88ba-4882-9908-f1e2f27c82ac" containerName="util" Nov 22 05:03:38 crc kubenswrapper[4818]: I1122 05:03:38.643772 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc4bf9a2-88ba-4882-9908-f1e2f27c82ac" containerName="util" Nov 22 05:03:38 crc kubenswrapper[4818]: E1122 05:03:38.643899 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc4bf9a2-88ba-4882-9908-f1e2f27c82ac" containerName="extract" Nov 22 05:03:38 crc kubenswrapper[4818]: I1122 05:03:38.644013 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc4bf9a2-88ba-4882-9908-f1e2f27c82ac" containerName="extract" Nov 22 05:03:38 crc kubenswrapper[4818]: E1122 05:03:38.644114 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc4bf9a2-88ba-4882-9908-f1e2f27c82ac" containerName="pull" Nov 22 05:03:38 crc kubenswrapper[4818]: I1122 05:03:38.644218 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc4bf9a2-88ba-4882-9908-f1e2f27c82ac" containerName="pull" Nov 22 05:03:38 crc kubenswrapper[4818]: I1122 05:03:38.644568 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc4bf9a2-88ba-4882-9908-f1e2f27c82ac" containerName="extract" Nov 22 05:03:38 crc kubenswrapper[4818]: I1122 05:03:38.645195 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-r2wjf" Nov 22 05:03:38 crc kubenswrapper[4818]: I1122 05:03:38.647120 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 22 05:03:38 crc kubenswrapper[4818]: I1122 05:03:38.647144 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-zfxzz" Nov 22 05:03:38 crc kubenswrapper[4818]: I1122 05:03:38.647288 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 22 05:03:38 crc kubenswrapper[4818]: I1122 05:03:38.655705 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-r2wjf"] Nov 22 05:03:38 crc kubenswrapper[4818]: I1122 05:03:38.808991 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfx4j\" (UniqueName: \"kubernetes.io/projected/05f359e3-f133-443a-9067-148ddc6fac3b-kube-api-access-xfx4j\") pod \"nmstate-operator-557fdffb88-r2wjf\" (UID: \"05f359e3-f133-443a-9067-148ddc6fac3b\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-r2wjf" Nov 22 05:03:38 crc kubenswrapper[4818]: I1122 05:03:38.910890 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfx4j\" (UniqueName: \"kubernetes.io/projected/05f359e3-f133-443a-9067-148ddc6fac3b-kube-api-access-xfx4j\") pod \"nmstate-operator-557fdffb88-r2wjf\" (UID: \"05f359e3-f133-443a-9067-148ddc6fac3b\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-r2wjf" Nov 22 05:03:38 crc kubenswrapper[4818]: I1122 05:03:38.938195 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfx4j\" (UniqueName: \"kubernetes.io/projected/05f359e3-f133-443a-9067-148ddc6fac3b-kube-api-access-xfx4j\") pod \"nmstate-operator-557fdffb88-r2wjf\" (UID: \"05f359e3-f133-443a-9067-148ddc6fac3b\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-r2wjf" Nov 22 05:03:38 crc kubenswrapper[4818]: I1122 05:03:38.967817 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-r2wjf" Nov 22 05:03:39 crc kubenswrapper[4818]: I1122 05:03:39.248829 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-r2wjf"] Nov 22 05:03:39 crc kubenswrapper[4818]: W1122 05:03:39.253404 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05f359e3_f133_443a_9067_148ddc6fac3b.slice/crio-2f1331acdfebb87466ba346ce10dc1c483620c57e751f30a923c642701ab97d1 WatchSource:0}: Error finding container 2f1331acdfebb87466ba346ce10dc1c483620c57e751f30a923c642701ab97d1: Status 404 returned error can't find the container with id 2f1331acdfebb87466ba346ce10dc1c483620c57e751f30a923c642701ab97d1 Nov 22 05:03:40 crc kubenswrapper[4818]: I1122 05:03:40.089524 4818 generic.go:334] "Generic (PLEG): container finished" podID="a988ba9e-742e-48e3-80de-0b7017216cf7" containerID="2f533eb71af4016865b414197b34b22d7630a1bd12854148fa5a2e0f60524f81" exitCode=0 Nov 22 05:03:40 crc kubenswrapper[4818]: I1122 05:03:40.089575 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msj6n" event={"ID":"a988ba9e-742e-48e3-80de-0b7017216cf7","Type":"ContainerDied","Data":"2f533eb71af4016865b414197b34b22d7630a1bd12854148fa5a2e0f60524f81"} Nov 22 05:03:40 crc kubenswrapper[4818]: I1122 05:03:40.090868 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-r2wjf" event={"ID":"05f359e3-f133-443a-9067-148ddc6fac3b","Type":"ContainerStarted","Data":"2f1331acdfebb87466ba346ce10dc1c483620c57e751f30a923c642701ab97d1"} Nov 22 05:03:41 crc kubenswrapper[4818]: I1122 05:03:41.102406 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msj6n" event={"ID":"a988ba9e-742e-48e3-80de-0b7017216cf7","Type":"ContainerStarted","Data":"0543d987f785c024f38d3fa621804e8de087278e83628414462497a56001793c"} Nov 22 05:03:41 crc kubenswrapper[4818]: I1122 05:03:41.128507 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-msj6n" podStartSLOduration=2.60881103 podStartE2EDuration="7.128471258s" podCreationTimestamp="2025-11-22 05:03:34 +0000 UTC" firstStartedPulling="2025-11-22 05:03:36.063395102 +0000 UTC m=+968.637811629" lastFinishedPulling="2025-11-22 05:03:40.58305533 +0000 UTC m=+973.157471857" observedRunningTime="2025-11-22 05:03:41.124437351 +0000 UTC m=+973.698853878" watchObservedRunningTime="2025-11-22 05:03:41.128471258 +0000 UTC m=+973.702887785" Nov 22 05:03:43 crc kubenswrapper[4818]: I1122 05:03:43.120159 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-r2wjf" event={"ID":"05f359e3-f133-443a-9067-148ddc6fac3b","Type":"ContainerStarted","Data":"3a83dfd2ce14738f4c71e047034a5b27bf3a2f324795734e0751afafae7658d7"} Nov 22 05:03:44 crc kubenswrapper[4818]: I1122 05:03:44.152330 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-r2wjf" podStartSLOduration=2.979587698 podStartE2EDuration="6.152313359s" podCreationTimestamp="2025-11-22 05:03:38 +0000 UTC" firstStartedPulling="2025-11-22 05:03:39.255959971 +0000 UTC m=+971.830376488" lastFinishedPulling="2025-11-22 05:03:42.428685572 +0000 UTC m=+975.003102149" observedRunningTime="2025-11-22 05:03:44.14932856 +0000 UTC m=+976.723745087" watchObservedRunningTime="2025-11-22 05:03:44.152313359 +0000 UTC m=+976.726729876" Nov 22 05:03:44 crc kubenswrapper[4818]: I1122 05:03:44.364233 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-msj6n" Nov 22 05:03:44 crc kubenswrapper[4818]: I1122 05:03:44.364308 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-msj6n" Nov 22 05:03:44 crc kubenswrapper[4818]: I1122 05:03:44.406786 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-msj6n" Nov 22 05:03:44 crc kubenswrapper[4818]: I1122 05:03:44.815369 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-trnm7" Nov 22 05:03:44 crc kubenswrapper[4818]: I1122 05:03:44.862064 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-trnm7" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.144945 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-f8v82"] Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.145818 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-f8v82" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.149510 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-nvwwz" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.154721 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-d94tc"] Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.155293 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-d94tc" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.158421 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.163274 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-ftnnk"] Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.163925 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-ftnnk" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.180741 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-d94tc"] Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.212470 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/695d974e-caa1-43c9-b619-363891b2bb7c-nmstate-lock\") pod \"nmstate-handler-ftnnk\" (UID: \"695d974e-caa1-43c9-b619-363891b2bb7c\") " pod="openshift-nmstate/nmstate-handler-ftnnk" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.212520 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdtvn\" (UniqueName: \"kubernetes.io/projected/695d974e-caa1-43c9-b619-363891b2bb7c-kube-api-access-mdtvn\") pod \"nmstate-handler-ftnnk\" (UID: \"695d974e-caa1-43c9-b619-363891b2bb7c\") " pod="openshift-nmstate/nmstate-handler-ftnnk" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.212559 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/695d974e-caa1-43c9-b619-363891b2bb7c-ovs-socket\") pod \"nmstate-handler-ftnnk\" (UID: \"695d974e-caa1-43c9-b619-363891b2bb7c\") " pod="openshift-nmstate/nmstate-handler-ftnnk" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.212576 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b0f0e1e5-3f4d-463b-8c00-91db5867fdfc-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-d94tc\" (UID: \"b0f0e1e5-3f4d-463b-8c00-91db5867fdfc\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-d94tc" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.212593 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckgqz\" (UniqueName: \"kubernetes.io/projected/b0f0e1e5-3f4d-463b-8c00-91db5867fdfc-kube-api-access-ckgqz\") pod \"nmstate-webhook-6b89b748d8-d94tc\" (UID: \"b0f0e1e5-3f4d-463b-8c00-91db5867fdfc\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-d94tc" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.212608 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xs98\" (UniqueName: \"kubernetes.io/projected/38740f9a-864c-44b3-8f95-f9ffecc8e9d5-kube-api-access-4xs98\") pod \"nmstate-metrics-5dcf9c57c5-f8v82\" (UID: \"38740f9a-864c-44b3-8f95-f9ffecc8e9d5\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-f8v82" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.212656 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/695d974e-caa1-43c9-b619-363891b2bb7c-dbus-socket\") pod \"nmstate-handler-ftnnk\" (UID: \"695d974e-caa1-43c9-b619-363891b2bb7c\") " pod="openshift-nmstate/nmstate-handler-ftnnk" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.220163 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-f8v82"] Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.223920 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-msj6n" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.288650 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-d2f5x"] Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.289370 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-d2f5x" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.292695 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.292781 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-4n55m" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.292920 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.299316 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-d2f5x"] Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.313287 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/695d974e-caa1-43c9-b619-363891b2bb7c-dbus-socket\") pod \"nmstate-handler-ftnnk\" (UID: \"695d974e-caa1-43c9-b619-363891b2bb7c\") " pod="openshift-nmstate/nmstate-handler-ftnnk" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.313331 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/614393c8-1594-4694-8281-a1e20cab9209-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-d2f5x\" (UID: \"614393c8-1594-4694-8281-a1e20cab9209\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-d2f5x" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.313362 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/695d974e-caa1-43c9-b619-363891b2bb7c-nmstate-lock\") pod \"nmstate-handler-ftnnk\" (UID: \"695d974e-caa1-43c9-b619-363891b2bb7c\") " pod="openshift-nmstate/nmstate-handler-ftnnk" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.313385 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/614393c8-1594-4694-8281-a1e20cab9209-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-d2f5x\" (UID: \"614393c8-1594-4694-8281-a1e20cab9209\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-d2f5x" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.313407 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdtvn\" (UniqueName: \"kubernetes.io/projected/695d974e-caa1-43c9-b619-363891b2bb7c-kube-api-access-mdtvn\") pod \"nmstate-handler-ftnnk\" (UID: \"695d974e-caa1-43c9-b619-363891b2bb7c\") " pod="openshift-nmstate/nmstate-handler-ftnnk" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.313430 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/695d974e-caa1-43c9-b619-363891b2bb7c-ovs-socket\") pod \"nmstate-handler-ftnnk\" (UID: \"695d974e-caa1-43c9-b619-363891b2bb7c\") " pod="openshift-nmstate/nmstate-handler-ftnnk" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.313447 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b0f0e1e5-3f4d-463b-8c00-91db5867fdfc-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-d94tc\" (UID: \"b0f0e1e5-3f4d-463b-8c00-91db5867fdfc\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-d94tc" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.313463 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckgqz\" (UniqueName: \"kubernetes.io/projected/b0f0e1e5-3f4d-463b-8c00-91db5867fdfc-kube-api-access-ckgqz\") pod \"nmstate-webhook-6b89b748d8-d94tc\" (UID: \"b0f0e1e5-3f4d-463b-8c00-91db5867fdfc\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-d94tc" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.313480 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xs98\" (UniqueName: \"kubernetes.io/projected/38740f9a-864c-44b3-8f95-f9ffecc8e9d5-kube-api-access-4xs98\") pod \"nmstate-metrics-5dcf9c57c5-f8v82\" (UID: \"38740f9a-864c-44b3-8f95-f9ffecc8e9d5\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-f8v82" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.313504 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qb66w\" (UniqueName: \"kubernetes.io/projected/614393c8-1594-4694-8281-a1e20cab9209-kube-api-access-qb66w\") pod \"nmstate-console-plugin-5874bd7bc5-d2f5x\" (UID: \"614393c8-1594-4694-8281-a1e20cab9209\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-d2f5x" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.313574 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/695d974e-caa1-43c9-b619-363891b2bb7c-nmstate-lock\") pod \"nmstate-handler-ftnnk\" (UID: \"695d974e-caa1-43c9-b619-363891b2bb7c\") " pod="openshift-nmstate/nmstate-handler-ftnnk" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.313667 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/695d974e-caa1-43c9-b619-363891b2bb7c-dbus-socket\") pod \"nmstate-handler-ftnnk\" (UID: \"695d974e-caa1-43c9-b619-363891b2bb7c\") " pod="openshift-nmstate/nmstate-handler-ftnnk" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.314100 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/695d974e-caa1-43c9-b619-363891b2bb7c-ovs-socket\") pod \"nmstate-handler-ftnnk\" (UID: \"695d974e-caa1-43c9-b619-363891b2bb7c\") " pod="openshift-nmstate/nmstate-handler-ftnnk" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.318534 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b0f0e1e5-3f4d-463b-8c00-91db5867fdfc-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-d94tc\" (UID: \"b0f0e1e5-3f4d-463b-8c00-91db5867fdfc\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-d94tc" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.329370 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdtvn\" (UniqueName: \"kubernetes.io/projected/695d974e-caa1-43c9-b619-363891b2bb7c-kube-api-access-mdtvn\") pod \"nmstate-handler-ftnnk\" (UID: \"695d974e-caa1-43c9-b619-363891b2bb7c\") " pod="openshift-nmstate/nmstate-handler-ftnnk" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.329452 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckgqz\" (UniqueName: \"kubernetes.io/projected/b0f0e1e5-3f4d-463b-8c00-91db5867fdfc-kube-api-access-ckgqz\") pod \"nmstate-webhook-6b89b748d8-d94tc\" (UID: \"b0f0e1e5-3f4d-463b-8c00-91db5867fdfc\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-d94tc" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.331310 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xs98\" (UniqueName: \"kubernetes.io/projected/38740f9a-864c-44b3-8f95-f9ffecc8e9d5-kube-api-access-4xs98\") pod \"nmstate-metrics-5dcf9c57c5-f8v82\" (UID: \"38740f9a-864c-44b3-8f95-f9ffecc8e9d5\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-f8v82" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.414344 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qb66w\" (UniqueName: \"kubernetes.io/projected/614393c8-1594-4694-8281-a1e20cab9209-kube-api-access-qb66w\") pod \"nmstate-console-plugin-5874bd7bc5-d2f5x\" (UID: \"614393c8-1594-4694-8281-a1e20cab9209\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-d2f5x" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.414598 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/614393c8-1594-4694-8281-a1e20cab9209-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-d2f5x\" (UID: \"614393c8-1594-4694-8281-a1e20cab9209\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-d2f5x" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.414644 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/614393c8-1594-4694-8281-a1e20cab9209-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-d2f5x\" (UID: \"614393c8-1594-4694-8281-a1e20cab9209\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-d2f5x" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.415544 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/614393c8-1594-4694-8281-a1e20cab9209-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-d2f5x\" (UID: \"614393c8-1594-4694-8281-a1e20cab9209\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-d2f5x" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.418135 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/614393c8-1594-4694-8281-a1e20cab9209-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-d2f5x\" (UID: \"614393c8-1594-4694-8281-a1e20cab9209\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-d2f5x" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.445744 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qb66w\" (UniqueName: \"kubernetes.io/projected/614393c8-1594-4694-8281-a1e20cab9209-kube-api-access-qb66w\") pod \"nmstate-console-plugin-5874bd7bc5-d2f5x\" (UID: \"614393c8-1594-4694-8281-a1e20cab9209\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-d2f5x" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.486735 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-f8v82" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.505916 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-d94tc" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.514324 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-ftnnk" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.547237 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7ddd6857cc-7cxk2"] Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.548669 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.554530 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7ddd6857cc-7cxk2"] Nov 22 05:03:45 crc kubenswrapper[4818]: W1122 05:03:45.580785 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod695d974e_caa1_43c9_b619_363891b2bb7c.slice/crio-9dfb51cdcb907612f524e51f0635cfe59356d859305490395e9c698dad91a074 WatchSource:0}: Error finding container 9dfb51cdcb907612f524e51f0635cfe59356d859305490395e9c698dad91a074: Status 404 returned error can't find the container with id 9dfb51cdcb907612f524e51f0635cfe59356d859305490395e9c698dad91a074 Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.610759 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-d2f5x" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.718453 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f4338e46-f595-4e2a-a279-f3a92ab43877-console-oauth-config\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.718498 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lxxg\" (UniqueName: \"kubernetes.io/projected/f4338e46-f595-4e2a-a279-f3a92ab43877-kube-api-access-2lxxg\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.718545 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f4338e46-f595-4e2a-a279-f3a92ab43877-oauth-serving-cert\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.718567 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4338e46-f595-4e2a-a279-f3a92ab43877-trusted-ca-bundle\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.718617 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f4338e46-f595-4e2a-a279-f3a92ab43877-console-config\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.718663 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f4338e46-f595-4e2a-a279-f3a92ab43877-service-ca\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.718704 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f4338e46-f595-4e2a-a279-f3a92ab43877-console-serving-cert\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.744446 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-d94tc"] Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.817555 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-f8v82"] Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.819643 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f4338e46-f595-4e2a-a279-f3a92ab43877-console-config\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.819705 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f4338e46-f595-4e2a-a279-f3a92ab43877-service-ca\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: W1122 05:03:45.819638 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38740f9a_864c_44b3_8f95_f9ffecc8e9d5.slice/crio-f7aab01dc17452dc271b140ade318a8aff0ea268f61b41f3b25919bd16175e3d WatchSource:0}: Error finding container f7aab01dc17452dc271b140ade318a8aff0ea268f61b41f3b25919bd16175e3d: Status 404 returned error can't find the container with id f7aab01dc17452dc271b140ade318a8aff0ea268f61b41f3b25919bd16175e3d Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.819751 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f4338e46-f595-4e2a-a279-f3a92ab43877-console-serving-cert\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.819784 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f4338e46-f595-4e2a-a279-f3a92ab43877-console-oauth-config\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.819807 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lxxg\" (UniqueName: \"kubernetes.io/projected/f4338e46-f595-4e2a-a279-f3a92ab43877-kube-api-access-2lxxg\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.819833 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f4338e46-f595-4e2a-a279-f3a92ab43877-oauth-serving-cert\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.819856 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4338e46-f595-4e2a-a279-f3a92ab43877-trusted-ca-bundle\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.820795 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f4338e46-f595-4e2a-a279-f3a92ab43877-console-config\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.821473 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f4338e46-f595-4e2a-a279-f3a92ab43877-service-ca\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.821929 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f4338e46-f595-4e2a-a279-f3a92ab43877-oauth-serving-cert\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.824243 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4338e46-f595-4e2a-a279-f3a92ab43877-trusted-ca-bundle\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.824524 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f4338e46-f595-4e2a-a279-f3a92ab43877-console-oauth-config\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.824661 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f4338e46-f595-4e2a-a279-f3a92ab43877-console-serving-cert\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.837900 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lxxg\" (UniqueName: \"kubernetes.io/projected/f4338e46-f595-4e2a-a279-f3a92ab43877-kube-api-access-2lxxg\") pod \"console-7ddd6857cc-7cxk2\" (UID: \"f4338e46-f595-4e2a-a279-f3a92ab43877\") " pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.853406 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-d2f5x"] Nov 22 05:03:45 crc kubenswrapper[4818]: W1122 05:03:45.854360 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod614393c8_1594_4694_8281_a1e20cab9209.slice/crio-29003c9d53d46a9ae994c5bbe743795e6befb0e8b44d721cb3b950af66cc77d6 WatchSource:0}: Error finding container 29003c9d53d46a9ae994c5bbe743795e6befb0e8b44d721cb3b950af66cc77d6: Status 404 returned error can't find the container with id 29003c9d53d46a9ae994c5bbe743795e6befb0e8b44d721cb3b950af66cc77d6 Nov 22 05:03:45 crc kubenswrapper[4818]: I1122 05:03:45.866321 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:46 crc kubenswrapper[4818]: I1122 05:03:46.035148 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7ddd6857cc-7cxk2"] Nov 22 05:03:46 crc kubenswrapper[4818]: W1122 05:03:46.040057 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4338e46_f595_4e2a_a279_f3a92ab43877.slice/crio-c679ef5cd851aeee45f710246d95825d9d2e562dc32be65b7468e218ab579adc WatchSource:0}: Error finding container c679ef5cd851aeee45f710246d95825d9d2e562dc32be65b7468e218ab579adc: Status 404 returned error can't find the container with id c679ef5cd851aeee45f710246d95825d9d2e562dc32be65b7468e218ab579adc Nov 22 05:03:46 crc kubenswrapper[4818]: I1122 05:03:46.137958 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-d94tc" event={"ID":"b0f0e1e5-3f4d-463b-8c00-91db5867fdfc","Type":"ContainerStarted","Data":"ad73287013ac96db1ee9ad41363ad26294425539c23dc79f817ac836fa8efec5"} Nov 22 05:03:46 crc kubenswrapper[4818]: I1122 05:03:46.138779 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-ftnnk" event={"ID":"695d974e-caa1-43c9-b619-363891b2bb7c","Type":"ContainerStarted","Data":"9dfb51cdcb907612f524e51f0635cfe59356d859305490395e9c698dad91a074"} Nov 22 05:03:46 crc kubenswrapper[4818]: I1122 05:03:46.139759 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7ddd6857cc-7cxk2" event={"ID":"f4338e46-f595-4e2a-a279-f3a92ab43877","Type":"ContainerStarted","Data":"c679ef5cd851aeee45f710246d95825d9d2e562dc32be65b7468e218ab579adc"} Nov 22 05:03:46 crc kubenswrapper[4818]: I1122 05:03:46.140634 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-d2f5x" event={"ID":"614393c8-1594-4694-8281-a1e20cab9209","Type":"ContainerStarted","Data":"29003c9d53d46a9ae994c5bbe743795e6befb0e8b44d721cb3b950af66cc77d6"} Nov 22 05:03:46 crc kubenswrapper[4818]: I1122 05:03:46.141588 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-f8v82" event={"ID":"38740f9a-864c-44b3-8f95-f9ffecc8e9d5","Type":"ContainerStarted","Data":"f7aab01dc17452dc271b140ade318a8aff0ea268f61b41f3b25919bd16175e3d"} Nov 22 05:03:47 crc kubenswrapper[4818]: I1122 05:03:47.150072 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7ddd6857cc-7cxk2" event={"ID":"f4338e46-f595-4e2a-a279-f3a92ab43877","Type":"ContainerStarted","Data":"5559a60ba7543c2fd794ecb890cd2739ff6fb0a76b0f55779ec50fa9e7954fd6"} Nov 22 05:03:47 crc kubenswrapper[4818]: I1122 05:03:47.170417 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7ddd6857cc-7cxk2" podStartSLOduration=2.170399909 podStartE2EDuration="2.170399909s" podCreationTimestamp="2025-11-22 05:03:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:03:47.166204987 +0000 UTC m=+979.740621514" watchObservedRunningTime="2025-11-22 05:03:47.170399909 +0000 UTC m=+979.744816436" Nov 22 05:03:47 crc kubenswrapper[4818]: I1122 05:03:47.629384 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-msj6n"] Nov 22 05:03:47 crc kubenswrapper[4818]: I1122 05:03:47.629660 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-msj6n" podUID="a988ba9e-742e-48e3-80de-0b7017216cf7" containerName="registry-server" containerID="cri-o://0543d987f785c024f38d3fa621804e8de087278e83628414462497a56001793c" gracePeriod=2 Nov 22 05:03:48 crc kubenswrapper[4818]: I1122 05:03:48.157454 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msj6n" event={"ID":"a988ba9e-742e-48e3-80de-0b7017216cf7","Type":"ContainerDied","Data":"0543d987f785c024f38d3fa621804e8de087278e83628414462497a56001793c"} Nov 22 05:03:48 crc kubenswrapper[4818]: I1122 05:03:48.157513 4818 generic.go:334] "Generic (PLEG): container finished" podID="a988ba9e-742e-48e3-80de-0b7017216cf7" containerID="0543d987f785c024f38d3fa621804e8de087278e83628414462497a56001793c" exitCode=0 Nov 22 05:03:48 crc kubenswrapper[4818]: I1122 05:03:48.432001 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-trnm7"] Nov 22 05:03:48 crc kubenswrapper[4818]: I1122 05:03:48.432766 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-trnm7" podUID="40cb0642-0b94-4089-bf50-f01484850cb5" containerName="registry-server" containerID="cri-o://6b0e799d2bc2e51a5e49890ae2b39c018723ac913f0f9be6926c9a064ea5581d" gracePeriod=2 Nov 22 05:03:49 crc kubenswrapper[4818]: I1122 05:03:49.166620 4818 generic.go:334] "Generic (PLEG): container finished" podID="40cb0642-0b94-4089-bf50-f01484850cb5" containerID="6b0e799d2bc2e51a5e49890ae2b39c018723ac913f0f9be6926c9a064ea5581d" exitCode=0 Nov 22 05:03:49 crc kubenswrapper[4818]: I1122 05:03:49.166692 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-trnm7" event={"ID":"40cb0642-0b94-4089-bf50-f01484850cb5","Type":"ContainerDied","Data":"6b0e799d2bc2e51a5e49890ae2b39c018723ac913f0f9be6926c9a064ea5581d"} Nov 22 05:03:49 crc kubenswrapper[4818]: I1122 05:03:49.931701 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-msj6n" Nov 22 05:03:50 crc kubenswrapper[4818]: I1122 05:03:50.082195 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a988ba9e-742e-48e3-80de-0b7017216cf7-catalog-content\") pod \"a988ba9e-742e-48e3-80de-0b7017216cf7\" (UID: \"a988ba9e-742e-48e3-80de-0b7017216cf7\") " Nov 22 05:03:50 crc kubenswrapper[4818]: I1122 05:03:50.082411 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfzj9\" (UniqueName: \"kubernetes.io/projected/a988ba9e-742e-48e3-80de-0b7017216cf7-kube-api-access-hfzj9\") pod \"a988ba9e-742e-48e3-80de-0b7017216cf7\" (UID: \"a988ba9e-742e-48e3-80de-0b7017216cf7\") " Nov 22 05:03:50 crc kubenswrapper[4818]: I1122 05:03:50.082459 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a988ba9e-742e-48e3-80de-0b7017216cf7-utilities\") pod \"a988ba9e-742e-48e3-80de-0b7017216cf7\" (UID: \"a988ba9e-742e-48e3-80de-0b7017216cf7\") " Nov 22 05:03:50 crc kubenswrapper[4818]: I1122 05:03:50.083997 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a988ba9e-742e-48e3-80de-0b7017216cf7-utilities" (OuterVolumeSpecName: "utilities") pod "a988ba9e-742e-48e3-80de-0b7017216cf7" (UID: "a988ba9e-742e-48e3-80de-0b7017216cf7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:03:50 crc kubenswrapper[4818]: I1122 05:03:50.084196 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a988ba9e-742e-48e3-80de-0b7017216cf7-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 05:03:50 crc kubenswrapper[4818]: I1122 05:03:50.088089 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a988ba9e-742e-48e3-80de-0b7017216cf7-kube-api-access-hfzj9" (OuterVolumeSpecName: "kube-api-access-hfzj9") pod "a988ba9e-742e-48e3-80de-0b7017216cf7" (UID: "a988ba9e-742e-48e3-80de-0b7017216cf7"). InnerVolumeSpecName "kube-api-access-hfzj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:03:50 crc kubenswrapper[4818]: I1122 05:03:50.103476 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a988ba9e-742e-48e3-80de-0b7017216cf7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a988ba9e-742e-48e3-80de-0b7017216cf7" (UID: "a988ba9e-742e-48e3-80de-0b7017216cf7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:03:50 crc kubenswrapper[4818]: I1122 05:03:50.179342 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msj6n" event={"ID":"a988ba9e-742e-48e3-80de-0b7017216cf7","Type":"ContainerDied","Data":"707ee3b2f285412ea3ef6d0ea6263ae48123c864edc1ff5be3806618e45f7110"} Nov 22 05:03:50 crc kubenswrapper[4818]: I1122 05:03:50.179404 4818 scope.go:117] "RemoveContainer" containerID="0543d987f785c024f38d3fa621804e8de087278e83628414462497a56001793c" Nov 22 05:03:50 crc kubenswrapper[4818]: I1122 05:03:50.179448 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-msj6n" Nov 22 05:03:50 crc kubenswrapper[4818]: I1122 05:03:50.184793 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a988ba9e-742e-48e3-80de-0b7017216cf7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 05:03:50 crc kubenswrapper[4818]: I1122 05:03:50.184821 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfzj9\" (UniqueName: \"kubernetes.io/projected/a988ba9e-742e-48e3-80de-0b7017216cf7-kube-api-access-hfzj9\") on node \"crc\" DevicePath \"\"" Nov 22 05:03:50 crc kubenswrapper[4818]: I1122 05:03:50.221544 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-msj6n"] Nov 22 05:03:50 crc kubenswrapper[4818]: I1122 05:03:50.226320 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-msj6n"] Nov 22 05:03:50 crc kubenswrapper[4818]: I1122 05:03:50.300001 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a988ba9e-742e-48e3-80de-0b7017216cf7" path="/var/lib/kubelet/pods/a988ba9e-742e-48e3-80de-0b7017216cf7/volumes" Nov 22 05:03:50 crc kubenswrapper[4818]: I1122 05:03:50.692139 4818 scope.go:117] "RemoveContainer" containerID="2f533eb71af4016865b414197b34b22d7630a1bd12854148fa5a2e0f60524f81" Nov 22 05:03:50 crc kubenswrapper[4818]: I1122 05:03:50.969477 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-trnm7" Nov 22 05:03:51 crc kubenswrapper[4818]: I1122 05:03:51.061898 4818 scope.go:117] "RemoveContainer" containerID="bc8a87598ee0971a39bd2a39323760f04170a5c9c908e26233b3c27694a02ea0" Nov 22 05:03:51 crc kubenswrapper[4818]: I1122 05:03:51.097307 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6td5\" (UniqueName: \"kubernetes.io/projected/40cb0642-0b94-4089-bf50-f01484850cb5-kube-api-access-p6td5\") pod \"40cb0642-0b94-4089-bf50-f01484850cb5\" (UID: \"40cb0642-0b94-4089-bf50-f01484850cb5\") " Nov 22 05:03:51 crc kubenswrapper[4818]: I1122 05:03:51.097351 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40cb0642-0b94-4089-bf50-f01484850cb5-catalog-content\") pod \"40cb0642-0b94-4089-bf50-f01484850cb5\" (UID: \"40cb0642-0b94-4089-bf50-f01484850cb5\") " Nov 22 05:03:51 crc kubenswrapper[4818]: I1122 05:03:51.101018 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40cb0642-0b94-4089-bf50-f01484850cb5-kube-api-access-p6td5" (OuterVolumeSpecName: "kube-api-access-p6td5") pod "40cb0642-0b94-4089-bf50-f01484850cb5" (UID: "40cb0642-0b94-4089-bf50-f01484850cb5"). InnerVolumeSpecName "kube-api-access-p6td5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:03:51 crc kubenswrapper[4818]: I1122 05:03:51.102419 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40cb0642-0b94-4089-bf50-f01484850cb5-utilities\") pod \"40cb0642-0b94-4089-bf50-f01484850cb5\" (UID: \"40cb0642-0b94-4089-bf50-f01484850cb5\") " Nov 22 05:03:51 crc kubenswrapper[4818]: I1122 05:03:51.103113 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6td5\" (UniqueName: \"kubernetes.io/projected/40cb0642-0b94-4089-bf50-f01484850cb5-kube-api-access-p6td5\") on node \"crc\" DevicePath \"\"" Nov 22 05:03:51 crc kubenswrapper[4818]: I1122 05:03:51.105382 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40cb0642-0b94-4089-bf50-f01484850cb5-utilities" (OuterVolumeSpecName: "utilities") pod "40cb0642-0b94-4089-bf50-f01484850cb5" (UID: "40cb0642-0b94-4089-bf50-f01484850cb5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:03:51 crc kubenswrapper[4818]: I1122 05:03:51.191360 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-trnm7" event={"ID":"40cb0642-0b94-4089-bf50-f01484850cb5","Type":"ContainerDied","Data":"cdbd90208b9d64fb76f051512d2520be1049040e45adfc7916c08ee40e8bd312"} Nov 22 05:03:51 crc kubenswrapper[4818]: I1122 05:03:51.191415 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-trnm7" Nov 22 05:03:51 crc kubenswrapper[4818]: I1122 05:03:51.191425 4818 scope.go:117] "RemoveContainer" containerID="6b0e799d2bc2e51a5e49890ae2b39c018723ac913f0f9be6926c9a064ea5581d" Nov 22 05:03:51 crc kubenswrapper[4818]: I1122 05:03:51.204097 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40cb0642-0b94-4089-bf50-f01484850cb5-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 05:03:51 crc kubenswrapper[4818]: I1122 05:03:51.217068 4818 scope.go:117] "RemoveContainer" containerID="45cf788a70df09717c4e968821098d6d01fa5de366704686190b23f2586f8d83" Nov 22 05:03:51 crc kubenswrapper[4818]: I1122 05:03:51.220233 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40cb0642-0b94-4089-bf50-f01484850cb5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "40cb0642-0b94-4089-bf50-f01484850cb5" (UID: "40cb0642-0b94-4089-bf50-f01484850cb5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:03:51 crc kubenswrapper[4818]: I1122 05:03:51.259572 4818 scope.go:117] "RemoveContainer" containerID="e9d1831e8d42c03e0f00d51874b3b74426511ada5b2e17189672f0d8cafd4f89" Nov 22 05:03:51 crc kubenswrapper[4818]: I1122 05:03:51.265855 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:03:51 crc kubenswrapper[4818]: I1122 05:03:51.265910 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:03:51 crc kubenswrapper[4818]: I1122 05:03:51.305383 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40cb0642-0b94-4089-bf50-f01484850cb5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 05:03:51 crc kubenswrapper[4818]: I1122 05:03:51.520648 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-trnm7"] Nov 22 05:03:51 crc kubenswrapper[4818]: I1122 05:03:51.523371 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-trnm7"] Nov 22 05:03:52 crc kubenswrapper[4818]: I1122 05:03:52.199708 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-f8v82" event={"ID":"38740f9a-864c-44b3-8f95-f9ffecc8e9d5","Type":"ContainerStarted","Data":"6de64c86c0c522fc296b251743df7bd7268a0bf29f881ce0423ac217ac07b87a"} Nov 22 05:03:52 crc kubenswrapper[4818]: I1122 05:03:52.201184 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-d94tc" event={"ID":"b0f0e1e5-3f4d-463b-8c00-91db5867fdfc","Type":"ContainerStarted","Data":"5bcd77266f063ae824c6981a9451dc77a22fc342f869511eaa8ae929289bfff9"} Nov 22 05:03:52 crc kubenswrapper[4818]: I1122 05:03:52.202059 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-d94tc" Nov 22 05:03:52 crc kubenswrapper[4818]: I1122 05:03:52.203778 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-ftnnk" event={"ID":"695d974e-caa1-43c9-b619-363891b2bb7c","Type":"ContainerStarted","Data":"207517060f65210447aa620eef7e0d3e0ee16bae6f7f8604d807cd6aa852b892"} Nov 22 05:03:52 crc kubenswrapper[4818]: I1122 05:03:52.204084 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-ftnnk" Nov 22 05:03:52 crc kubenswrapper[4818]: I1122 05:03:52.207392 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-d2f5x" event={"ID":"614393c8-1594-4694-8281-a1e20cab9209","Type":"ContainerStarted","Data":"ac5a26c2fe4efc105a151774422e77f6465b385fa38593d20998746d3f6442ee"} Nov 22 05:03:52 crc kubenswrapper[4818]: I1122 05:03:52.224830 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-d94tc" podStartSLOduration=1.89338367 podStartE2EDuration="7.224811321s" podCreationTimestamp="2025-11-22 05:03:45 +0000 UTC" firstStartedPulling="2025-11-22 05:03:45.759599492 +0000 UTC m=+978.334016019" lastFinishedPulling="2025-11-22 05:03:51.091027103 +0000 UTC m=+983.665443670" observedRunningTime="2025-11-22 05:03:52.218326188 +0000 UTC m=+984.792742725" watchObservedRunningTime="2025-11-22 05:03:52.224811321 +0000 UTC m=+984.799227848" Nov 22 05:03:52 crc kubenswrapper[4818]: I1122 05:03:52.238215 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-ftnnk" podStartSLOduration=1.735065088 podStartE2EDuration="7.238189956s" podCreationTimestamp="2025-11-22 05:03:45 +0000 UTC" firstStartedPulling="2025-11-22 05:03:45.587880185 +0000 UTC m=+978.162296722" lastFinishedPulling="2025-11-22 05:03:51.091005053 +0000 UTC m=+983.665421590" observedRunningTime="2025-11-22 05:03:52.232782462 +0000 UTC m=+984.807198989" watchObservedRunningTime="2025-11-22 05:03:52.238189956 +0000 UTC m=+984.812606493" Nov 22 05:03:52 crc kubenswrapper[4818]: I1122 05:03:52.260100 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-d2f5x" podStartSLOduration=2.026944692 podStartE2EDuration="7.260079129s" podCreationTimestamp="2025-11-22 05:03:45 +0000 UTC" firstStartedPulling="2025-11-22 05:03:45.856229862 +0000 UTC m=+978.430646389" lastFinishedPulling="2025-11-22 05:03:51.089364279 +0000 UTC m=+983.663780826" observedRunningTime="2025-11-22 05:03:52.248156391 +0000 UTC m=+984.822572918" watchObservedRunningTime="2025-11-22 05:03:52.260079129 +0000 UTC m=+984.834495656" Nov 22 05:03:52 crc kubenswrapper[4818]: I1122 05:03:52.297970 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40cb0642-0b94-4089-bf50-f01484850cb5" path="/var/lib/kubelet/pods/40cb0642-0b94-4089-bf50-f01484850cb5/volumes" Nov 22 05:03:55 crc kubenswrapper[4818]: I1122 05:03:55.236760 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-f8v82" event={"ID":"38740f9a-864c-44b3-8f95-f9ffecc8e9d5","Type":"ContainerStarted","Data":"3042475b0c0f403877f871a6c1a9f846f29e4b9a8b46faedc071bb72d33c56ec"} Nov 22 05:03:55 crc kubenswrapper[4818]: I1122 05:03:55.249869 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-f8v82" podStartSLOduration=2.009103059 podStartE2EDuration="10.249850344s" podCreationTimestamp="2025-11-22 05:03:45 +0000 UTC" firstStartedPulling="2025-11-22 05:03:45.821872979 +0000 UTC m=+978.396289506" lastFinishedPulling="2025-11-22 05:03:54.062620264 +0000 UTC m=+986.637036791" observedRunningTime="2025-11-22 05:03:55.248866748 +0000 UTC m=+987.823283295" watchObservedRunningTime="2025-11-22 05:03:55.249850344 +0000 UTC m=+987.824266871" Nov 22 05:03:55 crc kubenswrapper[4818]: I1122 05:03:55.866659 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:55 crc kubenswrapper[4818]: I1122 05:03:55.866971 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:55 crc kubenswrapper[4818]: I1122 05:03:55.871186 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:56 crc kubenswrapper[4818]: I1122 05:03:56.247035 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7ddd6857cc-7cxk2" Nov 22 05:03:56 crc kubenswrapper[4818]: I1122 05:03:56.290922 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-h56w2"] Nov 22 05:04:00 crc kubenswrapper[4818]: I1122 05:04:00.536942 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-ftnnk" Nov 22 05:04:05 crc kubenswrapper[4818]: I1122 05:04:05.514098 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-d94tc" Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.438014 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fbgl5"] Nov 22 05:04:13 crc kubenswrapper[4818]: E1122 05:04:13.438717 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40cb0642-0b94-4089-bf50-f01484850cb5" containerName="extract-content" Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.438731 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="40cb0642-0b94-4089-bf50-f01484850cb5" containerName="extract-content" Nov 22 05:04:13 crc kubenswrapper[4818]: E1122 05:04:13.438744 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a988ba9e-742e-48e3-80de-0b7017216cf7" containerName="registry-server" Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.438749 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a988ba9e-742e-48e3-80de-0b7017216cf7" containerName="registry-server" Nov 22 05:04:13 crc kubenswrapper[4818]: E1122 05:04:13.438758 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40cb0642-0b94-4089-bf50-f01484850cb5" containerName="extract-utilities" Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.438766 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="40cb0642-0b94-4089-bf50-f01484850cb5" containerName="extract-utilities" Nov 22 05:04:13 crc kubenswrapper[4818]: E1122 05:04:13.438780 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40cb0642-0b94-4089-bf50-f01484850cb5" containerName="registry-server" Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.438785 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="40cb0642-0b94-4089-bf50-f01484850cb5" containerName="registry-server" Nov 22 05:04:13 crc kubenswrapper[4818]: E1122 05:04:13.438797 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a988ba9e-742e-48e3-80de-0b7017216cf7" containerName="extract-content" Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.438802 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a988ba9e-742e-48e3-80de-0b7017216cf7" containerName="extract-content" Nov 22 05:04:13 crc kubenswrapper[4818]: E1122 05:04:13.438813 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a988ba9e-742e-48e3-80de-0b7017216cf7" containerName="extract-utilities" Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.438819 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a988ba9e-742e-48e3-80de-0b7017216cf7" containerName="extract-utilities" Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.438910 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="40cb0642-0b94-4089-bf50-f01484850cb5" containerName="registry-server" Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.438925 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a988ba9e-742e-48e3-80de-0b7017216cf7" containerName="registry-server" Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.439773 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fbgl5" Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.458189 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fbgl5"] Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.622405 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmdlp\" (UniqueName: \"kubernetes.io/projected/0ca92bf6-f49b-47be-81db-db6f3accbffd-kube-api-access-zmdlp\") pod \"certified-operators-fbgl5\" (UID: \"0ca92bf6-f49b-47be-81db-db6f3accbffd\") " pod="openshift-marketplace/certified-operators-fbgl5" Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.622561 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ca92bf6-f49b-47be-81db-db6f3accbffd-catalog-content\") pod \"certified-operators-fbgl5\" (UID: \"0ca92bf6-f49b-47be-81db-db6f3accbffd\") " pod="openshift-marketplace/certified-operators-fbgl5" Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.622612 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ca92bf6-f49b-47be-81db-db6f3accbffd-utilities\") pod \"certified-operators-fbgl5\" (UID: \"0ca92bf6-f49b-47be-81db-db6f3accbffd\") " pod="openshift-marketplace/certified-operators-fbgl5" Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.723590 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmdlp\" (UniqueName: \"kubernetes.io/projected/0ca92bf6-f49b-47be-81db-db6f3accbffd-kube-api-access-zmdlp\") pod \"certified-operators-fbgl5\" (UID: \"0ca92bf6-f49b-47be-81db-db6f3accbffd\") " pod="openshift-marketplace/certified-operators-fbgl5" Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.723688 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ca92bf6-f49b-47be-81db-db6f3accbffd-catalog-content\") pod \"certified-operators-fbgl5\" (UID: \"0ca92bf6-f49b-47be-81db-db6f3accbffd\") " pod="openshift-marketplace/certified-operators-fbgl5" Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.723717 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ca92bf6-f49b-47be-81db-db6f3accbffd-utilities\") pod \"certified-operators-fbgl5\" (UID: \"0ca92bf6-f49b-47be-81db-db6f3accbffd\") " pod="openshift-marketplace/certified-operators-fbgl5" Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.724419 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ca92bf6-f49b-47be-81db-db6f3accbffd-utilities\") pod \"certified-operators-fbgl5\" (UID: \"0ca92bf6-f49b-47be-81db-db6f3accbffd\") " pod="openshift-marketplace/certified-operators-fbgl5" Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.724504 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ca92bf6-f49b-47be-81db-db6f3accbffd-catalog-content\") pod \"certified-operators-fbgl5\" (UID: \"0ca92bf6-f49b-47be-81db-db6f3accbffd\") " pod="openshift-marketplace/certified-operators-fbgl5" Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.756111 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmdlp\" (UniqueName: \"kubernetes.io/projected/0ca92bf6-f49b-47be-81db-db6f3accbffd-kube-api-access-zmdlp\") pod \"certified-operators-fbgl5\" (UID: \"0ca92bf6-f49b-47be-81db-db6f3accbffd\") " pod="openshift-marketplace/certified-operators-fbgl5" Nov 22 05:04:13 crc kubenswrapper[4818]: I1122 05:04:13.758162 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fbgl5" Nov 22 05:04:14 crc kubenswrapper[4818]: I1122 05:04:14.074658 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fbgl5"] Nov 22 05:04:14 crc kubenswrapper[4818]: W1122 05:04:14.258890 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ca92bf6_f49b_47be_81db_db6f3accbffd.slice/crio-8d9b989422602d6f838c86ca560608b8593988add9c4224f25b85b20f43e34e7 WatchSource:0}: Error finding container 8d9b989422602d6f838c86ca560608b8593988add9c4224f25b85b20f43e34e7: Status 404 returned error can't find the container with id 8d9b989422602d6f838c86ca560608b8593988add9c4224f25b85b20f43e34e7 Nov 22 05:04:14 crc kubenswrapper[4818]: I1122 05:04:14.367379 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fbgl5" event={"ID":"0ca92bf6-f49b-47be-81db-db6f3accbffd","Type":"ContainerStarted","Data":"8d9b989422602d6f838c86ca560608b8593988add9c4224f25b85b20f43e34e7"} Nov 22 05:04:15 crc kubenswrapper[4818]: I1122 05:04:15.377140 4818 generic.go:334] "Generic (PLEG): container finished" podID="0ca92bf6-f49b-47be-81db-db6f3accbffd" containerID="de888651d8e21aad1490ec00b6dd9df0de7e9269ef1b2d49b517fe244192020c" exitCode=0 Nov 22 05:04:15 crc kubenswrapper[4818]: I1122 05:04:15.377317 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fbgl5" event={"ID":"0ca92bf6-f49b-47be-81db-db6f3accbffd","Type":"ContainerDied","Data":"de888651d8e21aad1490ec00b6dd9df0de7e9269ef1b2d49b517fe244192020c"} Nov 22 05:04:17 crc kubenswrapper[4818]: I1122 05:04:17.391046 4818 generic.go:334] "Generic (PLEG): container finished" podID="0ca92bf6-f49b-47be-81db-db6f3accbffd" containerID="b5401b9b54cd05a0019b83ef187aa6b4b897f3aef49b5eb64e1f389aba4ac992" exitCode=0 Nov 22 05:04:17 crc kubenswrapper[4818]: I1122 05:04:17.391141 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fbgl5" event={"ID":"0ca92bf6-f49b-47be-81db-db6f3accbffd","Type":"ContainerDied","Data":"b5401b9b54cd05a0019b83ef187aa6b4b897f3aef49b5eb64e1f389aba4ac992"} Nov 22 05:04:19 crc kubenswrapper[4818]: I1122 05:04:19.404613 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fbgl5" event={"ID":"0ca92bf6-f49b-47be-81db-db6f3accbffd","Type":"ContainerStarted","Data":"16c8d0e1cd3e26d503bb9ee821bfbd2eda31c65fbedbdf89f16952af1b3bbd8c"} Nov 22 05:04:19 crc kubenswrapper[4818]: I1122 05:04:19.423299 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fbgl5" podStartSLOduration=3.627623532 podStartE2EDuration="6.423285315s" podCreationTimestamp="2025-11-22 05:04:13 +0000 UTC" firstStartedPulling="2025-11-22 05:04:15.379639537 +0000 UTC m=+1007.954056084" lastFinishedPulling="2025-11-22 05:04:18.17530129 +0000 UTC m=+1010.749717867" observedRunningTime="2025-11-22 05:04:19.420732827 +0000 UTC m=+1011.995149394" watchObservedRunningTime="2025-11-22 05:04:19.423285315 +0000 UTC m=+1011.997701842" Nov 22 05:04:21 crc kubenswrapper[4818]: I1122 05:04:21.264799 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:04:21 crc kubenswrapper[4818]: I1122 05:04:21.265393 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:04:21 crc kubenswrapper[4818]: I1122 05:04:21.343955 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-h56w2" podUID="670a3117-320e-477c-8c4b-a54fe5a4ae9c" containerName="console" containerID="cri-o://41638750cf72db58a7884e847e359081b79c6ed5fa16f76b13d3e295dcb0fed8" gracePeriod=15 Nov 22 05:04:21 crc kubenswrapper[4818]: I1122 05:04:21.924638 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-h56w2_670a3117-320e-477c-8c4b-a54fe5a4ae9c/console/0.log" Nov 22 05:04:21 crc kubenswrapper[4818]: I1122 05:04:21.925003 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-h56w2" Nov 22 05:04:21 crc kubenswrapper[4818]: I1122 05:04:21.951798 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-trusted-ca-bundle\") pod \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " Nov 22 05:04:21 crc kubenswrapper[4818]: I1122 05:04:21.951845 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/670a3117-320e-477c-8c4b-a54fe5a4ae9c-console-oauth-config\") pod \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " Nov 22 05:04:21 crc kubenswrapper[4818]: I1122 05:04:21.951900 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-oauth-serving-cert\") pod \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " Nov 22 05:04:21 crc kubenswrapper[4818]: I1122 05:04:21.951964 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-service-ca\") pod \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " Nov 22 05:04:21 crc kubenswrapper[4818]: I1122 05:04:21.952011 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/670a3117-320e-477c-8c4b-a54fe5a4ae9c-console-serving-cert\") pod \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " Nov 22 05:04:21 crc kubenswrapper[4818]: I1122 05:04:21.952031 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8j7fh\" (UniqueName: \"kubernetes.io/projected/670a3117-320e-477c-8c4b-a54fe5a4ae9c-kube-api-access-8j7fh\") pod \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " Nov 22 05:04:21 crc kubenswrapper[4818]: I1122 05:04:21.952061 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-console-config\") pod \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\" (UID: \"670a3117-320e-477c-8c4b-a54fe5a4ae9c\") " Nov 22 05:04:21 crc kubenswrapper[4818]: I1122 05:04:21.952765 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "670a3117-320e-477c-8c4b-a54fe5a4ae9c" (UID: "670a3117-320e-477c-8c4b-a54fe5a4ae9c"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:04:21 crc kubenswrapper[4818]: I1122 05:04:21.952823 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-console-config" (OuterVolumeSpecName: "console-config") pod "670a3117-320e-477c-8c4b-a54fe5a4ae9c" (UID: "670a3117-320e-477c-8c4b-a54fe5a4ae9c"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:04:21 crc kubenswrapper[4818]: I1122 05:04:21.953309 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-service-ca" (OuterVolumeSpecName: "service-ca") pod "670a3117-320e-477c-8c4b-a54fe5a4ae9c" (UID: "670a3117-320e-477c-8c4b-a54fe5a4ae9c"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:04:21 crc kubenswrapper[4818]: I1122 05:04:21.953903 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "670a3117-320e-477c-8c4b-a54fe5a4ae9c" (UID: "670a3117-320e-477c-8c4b-a54fe5a4ae9c"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:04:21 crc kubenswrapper[4818]: I1122 05:04:21.960093 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/670a3117-320e-477c-8c4b-a54fe5a4ae9c-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "670a3117-320e-477c-8c4b-a54fe5a4ae9c" (UID: "670a3117-320e-477c-8c4b-a54fe5a4ae9c"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:04:21 crc kubenswrapper[4818]: I1122 05:04:21.960165 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/670a3117-320e-477c-8c4b-a54fe5a4ae9c-kube-api-access-8j7fh" (OuterVolumeSpecName: "kube-api-access-8j7fh") pod "670a3117-320e-477c-8c4b-a54fe5a4ae9c" (UID: "670a3117-320e-477c-8c4b-a54fe5a4ae9c"). InnerVolumeSpecName "kube-api-access-8j7fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:04:21 crc kubenswrapper[4818]: I1122 05:04:21.960539 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/670a3117-320e-477c-8c4b-a54fe5a4ae9c-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "670a3117-320e-477c-8c4b-a54fe5a4ae9c" (UID: "670a3117-320e-477c-8c4b-a54fe5a4ae9c"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.053220 4818 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.053473 4818 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.053483 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8j7fh\" (UniqueName: \"kubernetes.io/projected/670a3117-320e-477c-8c4b-a54fe5a4ae9c-kube-api-access-8j7fh\") on node \"crc\" DevicePath \"\"" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.053502 4818 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/670a3117-320e-477c-8c4b-a54fe5a4ae9c-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.053510 4818 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-console-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.053535 4818 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/670a3117-320e-477c-8c4b-a54fe5a4ae9c-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.053544 4818 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/670a3117-320e-477c-8c4b-a54fe5a4ae9c-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.426111 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-h56w2_670a3117-320e-477c-8c4b-a54fe5a4ae9c/console/0.log" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.426154 4818 generic.go:334] "Generic (PLEG): container finished" podID="670a3117-320e-477c-8c4b-a54fe5a4ae9c" containerID="41638750cf72db58a7884e847e359081b79c6ed5fa16f76b13d3e295dcb0fed8" exitCode=2 Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.426184 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-h56w2" event={"ID":"670a3117-320e-477c-8c4b-a54fe5a4ae9c","Type":"ContainerDied","Data":"41638750cf72db58a7884e847e359081b79c6ed5fa16f76b13d3e295dcb0fed8"} Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.426206 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-h56w2" event={"ID":"670a3117-320e-477c-8c4b-a54fe5a4ae9c","Type":"ContainerDied","Data":"2e51a4453f0a1e191a3e8883bc112a3d3f59a8d79ec7c58212962a629882a994"} Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.426225 4818 scope.go:117] "RemoveContainer" containerID="41638750cf72db58a7884e847e359081b79c6ed5fa16f76b13d3e295dcb0fed8" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.426351 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-h56w2" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.449869 4818 scope.go:117] "RemoveContainer" containerID="41638750cf72db58a7884e847e359081b79c6ed5fa16f76b13d3e295dcb0fed8" Nov 22 05:04:22 crc kubenswrapper[4818]: E1122 05:04:22.450710 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41638750cf72db58a7884e847e359081b79c6ed5fa16f76b13d3e295dcb0fed8\": container with ID starting with 41638750cf72db58a7884e847e359081b79c6ed5fa16f76b13d3e295dcb0fed8 not found: ID does not exist" containerID="41638750cf72db58a7884e847e359081b79c6ed5fa16f76b13d3e295dcb0fed8" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.450771 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41638750cf72db58a7884e847e359081b79c6ed5fa16f76b13d3e295dcb0fed8"} err="failed to get container status \"41638750cf72db58a7884e847e359081b79c6ed5fa16f76b13d3e295dcb0fed8\": rpc error: code = NotFound desc = could not find container \"41638750cf72db58a7884e847e359081b79c6ed5fa16f76b13d3e295dcb0fed8\": container with ID starting with 41638750cf72db58a7884e847e359081b79c6ed5fa16f76b13d3e295dcb0fed8 not found: ID does not exist" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.456284 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-h56w2"] Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.464412 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-h56w2"] Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.497217 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d"] Nov 22 05:04:22 crc kubenswrapper[4818]: E1122 05:04:22.497530 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="670a3117-320e-477c-8c4b-a54fe5a4ae9c" containerName="console" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.497547 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="670a3117-320e-477c-8c4b-a54fe5a4ae9c" containerName="console" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.497677 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="670a3117-320e-477c-8c4b-a54fe5a4ae9c" containerName="console" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.498503 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.500379 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.510978 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d"] Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.559079 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5dl8\" (UniqueName: \"kubernetes.io/projected/cd479145-6121-48b1-b54a-b122e24a03fd-kube-api-access-q5dl8\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d\" (UID: \"cd479145-6121-48b1-b54a-b122e24a03fd\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.559161 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cd479145-6121-48b1-b54a-b122e24a03fd-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d\" (UID: \"cd479145-6121-48b1-b54a-b122e24a03fd\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.559187 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cd479145-6121-48b1-b54a-b122e24a03fd-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d\" (UID: \"cd479145-6121-48b1-b54a-b122e24a03fd\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.660635 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5dl8\" (UniqueName: \"kubernetes.io/projected/cd479145-6121-48b1-b54a-b122e24a03fd-kube-api-access-q5dl8\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d\" (UID: \"cd479145-6121-48b1-b54a-b122e24a03fd\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.660757 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cd479145-6121-48b1-b54a-b122e24a03fd-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d\" (UID: \"cd479145-6121-48b1-b54a-b122e24a03fd\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.660801 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cd479145-6121-48b1-b54a-b122e24a03fd-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d\" (UID: \"cd479145-6121-48b1-b54a-b122e24a03fd\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.661800 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cd479145-6121-48b1-b54a-b122e24a03fd-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d\" (UID: \"cd479145-6121-48b1-b54a-b122e24a03fd\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.662019 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cd479145-6121-48b1-b54a-b122e24a03fd-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d\" (UID: \"cd479145-6121-48b1-b54a-b122e24a03fd\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.679873 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5dl8\" (UniqueName: \"kubernetes.io/projected/cd479145-6121-48b1-b54a-b122e24a03fd-kube-api-access-q5dl8\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d\" (UID: \"cd479145-6121-48b1-b54a-b122e24a03fd\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d" Nov 22 05:04:22 crc kubenswrapper[4818]: I1122 05:04:22.815364 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d" Nov 22 05:04:23 crc kubenswrapper[4818]: I1122 05:04:23.011771 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d"] Nov 22 05:04:23 crc kubenswrapper[4818]: W1122 05:04:23.018128 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd479145_6121_48b1_b54a_b122e24a03fd.slice/crio-d8d4fe87fd454905fa51fe532e5b7e1f2f54dca2d9fd6ad1d174f17b1f8f9f49 WatchSource:0}: Error finding container d8d4fe87fd454905fa51fe532e5b7e1f2f54dca2d9fd6ad1d174f17b1f8f9f49: Status 404 returned error can't find the container with id d8d4fe87fd454905fa51fe532e5b7e1f2f54dca2d9fd6ad1d174f17b1f8f9f49 Nov 22 05:04:23 crc kubenswrapper[4818]: I1122 05:04:23.434753 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d" event={"ID":"cd479145-6121-48b1-b54a-b122e24a03fd","Type":"ContainerStarted","Data":"a7369dab2c36c41eb33db611c066473a9040b24c5225049ee9d3974f1d596012"} Nov 22 05:04:23 crc kubenswrapper[4818]: I1122 05:04:23.434827 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d" event={"ID":"cd479145-6121-48b1-b54a-b122e24a03fd","Type":"ContainerStarted","Data":"d8d4fe87fd454905fa51fe532e5b7e1f2f54dca2d9fd6ad1d174f17b1f8f9f49"} Nov 22 05:04:23 crc kubenswrapper[4818]: I1122 05:04:23.758903 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fbgl5" Nov 22 05:04:23 crc kubenswrapper[4818]: I1122 05:04:23.759389 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fbgl5" Nov 22 05:04:23 crc kubenswrapper[4818]: I1122 05:04:23.798978 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fbgl5" Nov 22 05:04:24 crc kubenswrapper[4818]: I1122 05:04:24.298424 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="670a3117-320e-477c-8c4b-a54fe5a4ae9c" path="/var/lib/kubelet/pods/670a3117-320e-477c-8c4b-a54fe5a4ae9c/volumes" Nov 22 05:04:24 crc kubenswrapper[4818]: I1122 05:04:24.446982 4818 generic.go:334] "Generic (PLEG): container finished" podID="cd479145-6121-48b1-b54a-b122e24a03fd" containerID="a7369dab2c36c41eb33db611c066473a9040b24c5225049ee9d3974f1d596012" exitCode=0 Nov 22 05:04:24 crc kubenswrapper[4818]: I1122 05:04:24.447141 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d" event={"ID":"cd479145-6121-48b1-b54a-b122e24a03fd","Type":"ContainerDied","Data":"a7369dab2c36c41eb33db611c066473a9040b24c5225049ee9d3974f1d596012"} Nov 22 05:04:24 crc kubenswrapper[4818]: I1122 05:04:24.524732 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fbgl5" Nov 22 05:04:28 crc kubenswrapper[4818]: I1122 05:04:28.032224 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fbgl5"] Nov 22 05:04:28 crc kubenswrapper[4818]: I1122 05:04:28.032874 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fbgl5" podUID="0ca92bf6-f49b-47be-81db-db6f3accbffd" containerName="registry-server" containerID="cri-o://16c8d0e1cd3e26d503bb9ee821bfbd2eda31c65fbedbdf89f16952af1b3bbd8c" gracePeriod=2 Nov 22 05:04:30 crc kubenswrapper[4818]: I1122 05:04:30.492564 4818 generic.go:334] "Generic (PLEG): container finished" podID="0ca92bf6-f49b-47be-81db-db6f3accbffd" containerID="16c8d0e1cd3e26d503bb9ee821bfbd2eda31c65fbedbdf89f16952af1b3bbd8c" exitCode=0 Nov 22 05:04:30 crc kubenswrapper[4818]: I1122 05:04:30.492662 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fbgl5" event={"ID":"0ca92bf6-f49b-47be-81db-db6f3accbffd","Type":"ContainerDied","Data":"16c8d0e1cd3e26d503bb9ee821bfbd2eda31c65fbedbdf89f16952af1b3bbd8c"} Nov 22 05:04:32 crc kubenswrapper[4818]: I1122 05:04:32.094180 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fbgl5" Nov 22 05:04:32 crc kubenswrapper[4818]: I1122 05:04:32.187735 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ca92bf6-f49b-47be-81db-db6f3accbffd-catalog-content\") pod \"0ca92bf6-f49b-47be-81db-db6f3accbffd\" (UID: \"0ca92bf6-f49b-47be-81db-db6f3accbffd\") " Nov 22 05:04:32 crc kubenswrapper[4818]: I1122 05:04:32.187772 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmdlp\" (UniqueName: \"kubernetes.io/projected/0ca92bf6-f49b-47be-81db-db6f3accbffd-kube-api-access-zmdlp\") pod \"0ca92bf6-f49b-47be-81db-db6f3accbffd\" (UID: \"0ca92bf6-f49b-47be-81db-db6f3accbffd\") " Nov 22 05:04:32 crc kubenswrapper[4818]: I1122 05:04:32.187807 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ca92bf6-f49b-47be-81db-db6f3accbffd-utilities\") pod \"0ca92bf6-f49b-47be-81db-db6f3accbffd\" (UID: \"0ca92bf6-f49b-47be-81db-db6f3accbffd\") " Nov 22 05:04:32 crc kubenswrapper[4818]: I1122 05:04:32.188845 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ca92bf6-f49b-47be-81db-db6f3accbffd-utilities" (OuterVolumeSpecName: "utilities") pod "0ca92bf6-f49b-47be-81db-db6f3accbffd" (UID: "0ca92bf6-f49b-47be-81db-db6f3accbffd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:04:32 crc kubenswrapper[4818]: I1122 05:04:32.204007 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ca92bf6-f49b-47be-81db-db6f3accbffd-kube-api-access-zmdlp" (OuterVolumeSpecName: "kube-api-access-zmdlp") pod "0ca92bf6-f49b-47be-81db-db6f3accbffd" (UID: "0ca92bf6-f49b-47be-81db-db6f3accbffd"). InnerVolumeSpecName "kube-api-access-zmdlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:04:32 crc kubenswrapper[4818]: I1122 05:04:32.268116 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ca92bf6-f49b-47be-81db-db6f3accbffd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ca92bf6-f49b-47be-81db-db6f3accbffd" (UID: "0ca92bf6-f49b-47be-81db-db6f3accbffd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:04:32 crc kubenswrapper[4818]: I1122 05:04:32.289283 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ca92bf6-f49b-47be-81db-db6f3accbffd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 05:04:32 crc kubenswrapper[4818]: I1122 05:04:32.289326 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmdlp\" (UniqueName: \"kubernetes.io/projected/0ca92bf6-f49b-47be-81db-db6f3accbffd-kube-api-access-zmdlp\") on node \"crc\" DevicePath \"\"" Nov 22 05:04:32 crc kubenswrapper[4818]: I1122 05:04:32.289341 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ca92bf6-f49b-47be-81db-db6f3accbffd-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 05:04:32 crc kubenswrapper[4818]: I1122 05:04:32.507490 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fbgl5" event={"ID":"0ca92bf6-f49b-47be-81db-db6f3accbffd","Type":"ContainerDied","Data":"8d9b989422602d6f838c86ca560608b8593988add9c4224f25b85b20f43e34e7"} Nov 22 05:04:32 crc kubenswrapper[4818]: I1122 05:04:32.507537 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fbgl5" Nov 22 05:04:32 crc kubenswrapper[4818]: I1122 05:04:32.507551 4818 scope.go:117] "RemoveContainer" containerID="16c8d0e1cd3e26d503bb9ee821bfbd2eda31c65fbedbdf89f16952af1b3bbd8c" Nov 22 05:04:32 crc kubenswrapper[4818]: I1122 05:04:32.511391 4818 generic.go:334] "Generic (PLEG): container finished" podID="cd479145-6121-48b1-b54a-b122e24a03fd" containerID="6a78c1461e38cb7bced68e0067c3a7224fff310b1bc2c27422020640e94ddad4" exitCode=0 Nov 22 05:04:32 crc kubenswrapper[4818]: I1122 05:04:32.511438 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d" event={"ID":"cd479145-6121-48b1-b54a-b122e24a03fd","Type":"ContainerDied","Data":"6a78c1461e38cb7bced68e0067c3a7224fff310b1bc2c27422020640e94ddad4"} Nov 22 05:04:32 crc kubenswrapper[4818]: I1122 05:04:32.580179 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fbgl5"] Nov 22 05:04:32 crc kubenswrapper[4818]: I1122 05:04:32.580501 4818 scope.go:117] "RemoveContainer" containerID="b5401b9b54cd05a0019b83ef187aa6b4b897f3aef49b5eb64e1f389aba4ac992" Nov 22 05:04:32 crc kubenswrapper[4818]: I1122 05:04:32.588245 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fbgl5"] Nov 22 05:04:32 crc kubenswrapper[4818]: I1122 05:04:32.603369 4818 scope.go:117] "RemoveContainer" containerID="de888651d8e21aad1490ec00b6dd9df0de7e9269ef1b2d49b517fe244192020c" Nov 22 05:04:33 crc kubenswrapper[4818]: I1122 05:04:33.523623 4818 generic.go:334] "Generic (PLEG): container finished" podID="cd479145-6121-48b1-b54a-b122e24a03fd" containerID="39de8915cec9f865582ef3ccde54514f2a01b8f9b7cdb524b9357c7c47a499a0" exitCode=0 Nov 22 05:04:33 crc kubenswrapper[4818]: I1122 05:04:33.523758 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d" event={"ID":"cd479145-6121-48b1-b54a-b122e24a03fd","Type":"ContainerDied","Data":"39de8915cec9f865582ef3ccde54514f2a01b8f9b7cdb524b9357c7c47a499a0"} Nov 22 05:04:34 crc kubenswrapper[4818]: I1122 05:04:34.303629 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ca92bf6-f49b-47be-81db-db6f3accbffd" path="/var/lib/kubelet/pods/0ca92bf6-f49b-47be-81db-db6f3accbffd/volumes" Nov 22 05:04:34 crc kubenswrapper[4818]: I1122 05:04:34.796877 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d" Nov 22 05:04:34 crc kubenswrapper[4818]: I1122 05:04:34.925482 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cd479145-6121-48b1-b54a-b122e24a03fd-util\") pod \"cd479145-6121-48b1-b54a-b122e24a03fd\" (UID: \"cd479145-6121-48b1-b54a-b122e24a03fd\") " Nov 22 05:04:34 crc kubenswrapper[4818]: I1122 05:04:34.925619 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cd479145-6121-48b1-b54a-b122e24a03fd-bundle\") pod \"cd479145-6121-48b1-b54a-b122e24a03fd\" (UID: \"cd479145-6121-48b1-b54a-b122e24a03fd\") " Nov 22 05:04:34 crc kubenswrapper[4818]: I1122 05:04:34.925702 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5dl8\" (UniqueName: \"kubernetes.io/projected/cd479145-6121-48b1-b54a-b122e24a03fd-kube-api-access-q5dl8\") pod \"cd479145-6121-48b1-b54a-b122e24a03fd\" (UID: \"cd479145-6121-48b1-b54a-b122e24a03fd\") " Nov 22 05:04:34 crc kubenswrapper[4818]: I1122 05:04:34.927134 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd479145-6121-48b1-b54a-b122e24a03fd-bundle" (OuterVolumeSpecName: "bundle") pod "cd479145-6121-48b1-b54a-b122e24a03fd" (UID: "cd479145-6121-48b1-b54a-b122e24a03fd"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:04:34 crc kubenswrapper[4818]: I1122 05:04:34.932229 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd479145-6121-48b1-b54a-b122e24a03fd-kube-api-access-q5dl8" (OuterVolumeSpecName: "kube-api-access-q5dl8") pod "cd479145-6121-48b1-b54a-b122e24a03fd" (UID: "cd479145-6121-48b1-b54a-b122e24a03fd"). InnerVolumeSpecName "kube-api-access-q5dl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:04:34 crc kubenswrapper[4818]: I1122 05:04:34.938582 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd479145-6121-48b1-b54a-b122e24a03fd-util" (OuterVolumeSpecName: "util") pod "cd479145-6121-48b1-b54a-b122e24a03fd" (UID: "cd479145-6121-48b1-b54a-b122e24a03fd"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:04:35 crc kubenswrapper[4818]: I1122 05:04:35.027392 4818 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cd479145-6121-48b1-b54a-b122e24a03fd-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:04:35 crc kubenswrapper[4818]: I1122 05:04:35.027422 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5dl8\" (UniqueName: \"kubernetes.io/projected/cd479145-6121-48b1-b54a-b122e24a03fd-kube-api-access-q5dl8\") on node \"crc\" DevicePath \"\"" Nov 22 05:04:35 crc kubenswrapper[4818]: I1122 05:04:35.027432 4818 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cd479145-6121-48b1-b54a-b122e24a03fd-util\") on node \"crc\" DevicePath \"\"" Nov 22 05:04:35 crc kubenswrapper[4818]: I1122 05:04:35.536077 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d" event={"ID":"cd479145-6121-48b1-b54a-b122e24a03fd","Type":"ContainerDied","Data":"d8d4fe87fd454905fa51fe532e5b7e1f2f54dca2d9fd6ad1d174f17b1f8f9f49"} Nov 22 05:04:35 crc kubenswrapper[4818]: I1122 05:04:35.536115 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8d4fe87fd454905fa51fe532e5b7e1f2f54dca2d9fd6ad1d174f17b1f8f9f49" Nov 22 05:04:35 crc kubenswrapper[4818]: I1122 05:04:35.536169 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.036148 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-94k9s"] Nov 22 05:04:46 crc kubenswrapper[4818]: E1122 05:04:46.036889 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca92bf6-f49b-47be-81db-db6f3accbffd" containerName="extract-content" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.036903 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca92bf6-f49b-47be-81db-db6f3accbffd" containerName="extract-content" Nov 22 05:04:46 crc kubenswrapper[4818]: E1122 05:04:46.036918 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd479145-6121-48b1-b54a-b122e24a03fd" containerName="pull" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.036925 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd479145-6121-48b1-b54a-b122e24a03fd" containerName="pull" Nov 22 05:04:46 crc kubenswrapper[4818]: E1122 05:04:46.036938 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca92bf6-f49b-47be-81db-db6f3accbffd" containerName="extract-utilities" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.036946 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca92bf6-f49b-47be-81db-db6f3accbffd" containerName="extract-utilities" Nov 22 05:04:46 crc kubenswrapper[4818]: E1122 05:04:46.036957 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca92bf6-f49b-47be-81db-db6f3accbffd" containerName="registry-server" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.036964 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca92bf6-f49b-47be-81db-db6f3accbffd" containerName="registry-server" Nov 22 05:04:46 crc kubenswrapper[4818]: E1122 05:04:46.036982 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd479145-6121-48b1-b54a-b122e24a03fd" containerName="util" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.036989 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd479145-6121-48b1-b54a-b122e24a03fd" containerName="util" Nov 22 05:04:46 crc kubenswrapper[4818]: E1122 05:04:46.037004 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd479145-6121-48b1-b54a-b122e24a03fd" containerName="extract" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.037011 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd479145-6121-48b1-b54a-b122e24a03fd" containerName="extract" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.037125 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd479145-6121-48b1-b54a-b122e24a03fd" containerName="extract" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.037136 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ca92bf6-f49b-47be-81db-db6f3accbffd" containerName="registry-server" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.038079 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-94k9s" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.051505 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-94k9s"] Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.157345 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64fdc198-7025-4a93-903f-0a54a1f80967-catalog-content\") pod \"community-operators-94k9s\" (UID: \"64fdc198-7025-4a93-903f-0a54a1f80967\") " pod="openshift-marketplace/community-operators-94k9s" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.157423 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64fdc198-7025-4a93-903f-0a54a1f80967-utilities\") pod \"community-operators-94k9s\" (UID: \"64fdc198-7025-4a93-903f-0a54a1f80967\") " pod="openshift-marketplace/community-operators-94k9s" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.157479 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8zbv\" (UniqueName: \"kubernetes.io/projected/64fdc198-7025-4a93-903f-0a54a1f80967-kube-api-access-k8zbv\") pod \"community-operators-94k9s\" (UID: \"64fdc198-7025-4a93-903f-0a54a1f80967\") " pod="openshift-marketplace/community-operators-94k9s" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.258818 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8zbv\" (UniqueName: \"kubernetes.io/projected/64fdc198-7025-4a93-903f-0a54a1f80967-kube-api-access-k8zbv\") pod \"community-operators-94k9s\" (UID: \"64fdc198-7025-4a93-903f-0a54a1f80967\") " pod="openshift-marketplace/community-operators-94k9s" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.258936 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64fdc198-7025-4a93-903f-0a54a1f80967-catalog-content\") pod \"community-operators-94k9s\" (UID: \"64fdc198-7025-4a93-903f-0a54a1f80967\") " pod="openshift-marketplace/community-operators-94k9s" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.258971 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64fdc198-7025-4a93-903f-0a54a1f80967-utilities\") pod \"community-operators-94k9s\" (UID: \"64fdc198-7025-4a93-903f-0a54a1f80967\") " pod="openshift-marketplace/community-operators-94k9s" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.259514 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64fdc198-7025-4a93-903f-0a54a1f80967-utilities\") pod \"community-operators-94k9s\" (UID: \"64fdc198-7025-4a93-903f-0a54a1f80967\") " pod="openshift-marketplace/community-operators-94k9s" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.259666 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64fdc198-7025-4a93-903f-0a54a1f80967-catalog-content\") pod \"community-operators-94k9s\" (UID: \"64fdc198-7025-4a93-903f-0a54a1f80967\") " pod="openshift-marketplace/community-operators-94k9s" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.283093 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8zbv\" (UniqueName: \"kubernetes.io/projected/64fdc198-7025-4a93-903f-0a54a1f80967-kube-api-access-k8zbv\") pod \"community-operators-94k9s\" (UID: \"64fdc198-7025-4a93-903f-0a54a1f80967\") " pod="openshift-marketplace/community-operators-94k9s" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.352632 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-94k9s" Nov 22 05:04:46 crc kubenswrapper[4818]: I1122 05:04:46.800868 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-94k9s"] Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.302939 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b7cbb7cc5-ff54p"] Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.303971 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5b7cbb7cc5-ff54p" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.306703 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.307057 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.307208 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.307236 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-24k6n" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.307647 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.320484 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b7cbb7cc5-ff54p"] Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.374730 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qxml\" (UniqueName: \"kubernetes.io/projected/254a82ef-b456-4309-93fb-a236494bf3d4-kube-api-access-2qxml\") pod \"metallb-operator-controller-manager-5b7cbb7cc5-ff54p\" (UID: \"254a82ef-b456-4309-93fb-a236494bf3d4\") " pod="metallb-system/metallb-operator-controller-manager-5b7cbb7cc5-ff54p" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.374779 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/254a82ef-b456-4309-93fb-a236494bf3d4-apiservice-cert\") pod \"metallb-operator-controller-manager-5b7cbb7cc5-ff54p\" (UID: \"254a82ef-b456-4309-93fb-a236494bf3d4\") " pod="metallb-system/metallb-operator-controller-manager-5b7cbb7cc5-ff54p" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.374816 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/254a82ef-b456-4309-93fb-a236494bf3d4-webhook-cert\") pod \"metallb-operator-controller-manager-5b7cbb7cc5-ff54p\" (UID: \"254a82ef-b456-4309-93fb-a236494bf3d4\") " pod="metallb-system/metallb-operator-controller-manager-5b7cbb7cc5-ff54p" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.476367 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qxml\" (UniqueName: \"kubernetes.io/projected/254a82ef-b456-4309-93fb-a236494bf3d4-kube-api-access-2qxml\") pod \"metallb-operator-controller-manager-5b7cbb7cc5-ff54p\" (UID: \"254a82ef-b456-4309-93fb-a236494bf3d4\") " pod="metallb-system/metallb-operator-controller-manager-5b7cbb7cc5-ff54p" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.476421 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/254a82ef-b456-4309-93fb-a236494bf3d4-apiservice-cert\") pod \"metallb-operator-controller-manager-5b7cbb7cc5-ff54p\" (UID: \"254a82ef-b456-4309-93fb-a236494bf3d4\") " pod="metallb-system/metallb-operator-controller-manager-5b7cbb7cc5-ff54p" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.476463 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/254a82ef-b456-4309-93fb-a236494bf3d4-webhook-cert\") pod \"metallb-operator-controller-manager-5b7cbb7cc5-ff54p\" (UID: \"254a82ef-b456-4309-93fb-a236494bf3d4\") " pod="metallb-system/metallb-operator-controller-manager-5b7cbb7cc5-ff54p" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.483206 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/254a82ef-b456-4309-93fb-a236494bf3d4-webhook-cert\") pod \"metallb-operator-controller-manager-5b7cbb7cc5-ff54p\" (UID: \"254a82ef-b456-4309-93fb-a236494bf3d4\") " pod="metallb-system/metallb-operator-controller-manager-5b7cbb7cc5-ff54p" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.483233 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/254a82ef-b456-4309-93fb-a236494bf3d4-apiservice-cert\") pod \"metallb-operator-controller-manager-5b7cbb7cc5-ff54p\" (UID: \"254a82ef-b456-4309-93fb-a236494bf3d4\") " pod="metallb-system/metallb-operator-controller-manager-5b7cbb7cc5-ff54p" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.497014 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qxml\" (UniqueName: \"kubernetes.io/projected/254a82ef-b456-4309-93fb-a236494bf3d4-kube-api-access-2qxml\") pod \"metallb-operator-controller-manager-5b7cbb7cc5-ff54p\" (UID: \"254a82ef-b456-4309-93fb-a236494bf3d4\") " pod="metallb-system/metallb-operator-controller-manager-5b7cbb7cc5-ff54p" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.542533 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-757d8876f9-6qjkv"] Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.543147 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-757d8876f9-6qjkv" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.545649 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.545677 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.545798 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-h5gsm" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.558821 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-757d8876f9-6qjkv"] Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.577482 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/63241779-80e0-43a3-bb9c-6530a4a6ae05-webhook-cert\") pod \"metallb-operator-webhook-server-757d8876f9-6qjkv\" (UID: \"63241779-80e0-43a3-bb9c-6530a4a6ae05\") " pod="metallb-system/metallb-operator-webhook-server-757d8876f9-6qjkv" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.577556 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flld4\" (UniqueName: \"kubernetes.io/projected/63241779-80e0-43a3-bb9c-6530a4a6ae05-kube-api-access-flld4\") pod \"metallb-operator-webhook-server-757d8876f9-6qjkv\" (UID: \"63241779-80e0-43a3-bb9c-6530a4a6ae05\") " pod="metallb-system/metallb-operator-webhook-server-757d8876f9-6qjkv" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.577618 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/63241779-80e0-43a3-bb9c-6530a4a6ae05-apiservice-cert\") pod \"metallb-operator-webhook-server-757d8876f9-6qjkv\" (UID: \"63241779-80e0-43a3-bb9c-6530a4a6ae05\") " pod="metallb-system/metallb-operator-webhook-server-757d8876f9-6qjkv" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.607689 4818 generic.go:334] "Generic (PLEG): container finished" podID="64fdc198-7025-4a93-903f-0a54a1f80967" containerID="4755e538b235b383b48be6b85296cfd969bae074dbac2ea37928a22d8d3d551f" exitCode=0 Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.607733 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-94k9s" event={"ID":"64fdc198-7025-4a93-903f-0a54a1f80967","Type":"ContainerDied","Data":"4755e538b235b383b48be6b85296cfd969bae074dbac2ea37928a22d8d3d551f"} Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.607764 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-94k9s" event={"ID":"64fdc198-7025-4a93-903f-0a54a1f80967","Type":"ContainerStarted","Data":"a380a2972105f8755c98a45bd44044d39c18de76d676dee2b87c2006915e5927"} Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.621453 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5b7cbb7cc5-ff54p" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.679608 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/63241779-80e0-43a3-bb9c-6530a4a6ae05-webhook-cert\") pod \"metallb-operator-webhook-server-757d8876f9-6qjkv\" (UID: \"63241779-80e0-43a3-bb9c-6530a4a6ae05\") " pod="metallb-system/metallb-operator-webhook-server-757d8876f9-6qjkv" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.679685 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flld4\" (UniqueName: \"kubernetes.io/projected/63241779-80e0-43a3-bb9c-6530a4a6ae05-kube-api-access-flld4\") pod \"metallb-operator-webhook-server-757d8876f9-6qjkv\" (UID: \"63241779-80e0-43a3-bb9c-6530a4a6ae05\") " pod="metallb-system/metallb-operator-webhook-server-757d8876f9-6qjkv" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.679752 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/63241779-80e0-43a3-bb9c-6530a4a6ae05-apiservice-cert\") pod \"metallb-operator-webhook-server-757d8876f9-6qjkv\" (UID: \"63241779-80e0-43a3-bb9c-6530a4a6ae05\") " pod="metallb-system/metallb-operator-webhook-server-757d8876f9-6qjkv" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.685734 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/63241779-80e0-43a3-bb9c-6530a4a6ae05-webhook-cert\") pod \"metallb-operator-webhook-server-757d8876f9-6qjkv\" (UID: \"63241779-80e0-43a3-bb9c-6530a4a6ae05\") " pod="metallb-system/metallb-operator-webhook-server-757d8876f9-6qjkv" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.685801 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/63241779-80e0-43a3-bb9c-6530a4a6ae05-apiservice-cert\") pod \"metallb-operator-webhook-server-757d8876f9-6qjkv\" (UID: \"63241779-80e0-43a3-bb9c-6530a4a6ae05\") " pod="metallb-system/metallb-operator-webhook-server-757d8876f9-6qjkv" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.706036 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flld4\" (UniqueName: \"kubernetes.io/projected/63241779-80e0-43a3-bb9c-6530a4a6ae05-kube-api-access-flld4\") pod \"metallb-operator-webhook-server-757d8876f9-6qjkv\" (UID: \"63241779-80e0-43a3-bb9c-6530a4a6ae05\") " pod="metallb-system/metallb-operator-webhook-server-757d8876f9-6qjkv" Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.867531 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b7cbb7cc5-ff54p"] Nov 22 05:04:47 crc kubenswrapper[4818]: I1122 05:04:47.892668 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-757d8876f9-6qjkv" Nov 22 05:04:48 crc kubenswrapper[4818]: I1122 05:04:48.070894 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-757d8876f9-6qjkv"] Nov 22 05:04:48 crc kubenswrapper[4818]: W1122 05:04:48.078997 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63241779_80e0_43a3_bb9c_6530a4a6ae05.slice/crio-90e90d8f3a1df72f954731fd0f495a1edf293606e5d5c7156d95566cb51e2d6c WatchSource:0}: Error finding container 90e90d8f3a1df72f954731fd0f495a1edf293606e5d5c7156d95566cb51e2d6c: Status 404 returned error can't find the container with id 90e90d8f3a1df72f954731fd0f495a1edf293606e5d5c7156d95566cb51e2d6c Nov 22 05:04:48 crc kubenswrapper[4818]: I1122 05:04:48.614145 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-757d8876f9-6qjkv" event={"ID":"63241779-80e0-43a3-bb9c-6530a4a6ae05","Type":"ContainerStarted","Data":"90e90d8f3a1df72f954731fd0f495a1edf293606e5d5c7156d95566cb51e2d6c"} Nov 22 05:04:48 crc kubenswrapper[4818]: I1122 05:04:48.615355 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5b7cbb7cc5-ff54p" event={"ID":"254a82ef-b456-4309-93fb-a236494bf3d4","Type":"ContainerStarted","Data":"3e704298d05342f212d5d14a2196e15a022376c3a1cb00e3393792f7a0db0d38"} Nov 22 05:04:49 crc kubenswrapper[4818]: I1122 05:04:49.625040 4818 generic.go:334] "Generic (PLEG): container finished" podID="64fdc198-7025-4a93-903f-0a54a1f80967" containerID="b87d2e5f6d106f7819d5241723b71cb7418d1fd7508f9960afea370e7f14cd5e" exitCode=0 Nov 22 05:04:49 crc kubenswrapper[4818]: I1122 05:04:49.625336 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-94k9s" event={"ID":"64fdc198-7025-4a93-903f-0a54a1f80967","Type":"ContainerDied","Data":"b87d2e5f6d106f7819d5241723b71cb7418d1fd7508f9960afea370e7f14cd5e"} Nov 22 05:04:50 crc kubenswrapper[4818]: I1122 05:04:50.641989 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-94k9s" event={"ID":"64fdc198-7025-4a93-903f-0a54a1f80967","Type":"ContainerStarted","Data":"6e8d8dcf70a4b8b8921705a1c4e6e8b4614f7c827d480b36c185afc49ab61fe4"} Nov 22 05:04:50 crc kubenswrapper[4818]: I1122 05:04:50.667494 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-94k9s" podStartSLOduration=2.230199743 podStartE2EDuration="4.667475752s" podCreationTimestamp="2025-11-22 05:04:46 +0000 UTC" firstStartedPulling="2025-11-22 05:04:47.609533512 +0000 UTC m=+1040.183950039" lastFinishedPulling="2025-11-22 05:04:50.046809521 +0000 UTC m=+1042.621226048" observedRunningTime="2025-11-22 05:04:50.666777783 +0000 UTC m=+1043.241194310" watchObservedRunningTime="2025-11-22 05:04:50.667475752 +0000 UTC m=+1043.241892279" Nov 22 05:04:51 crc kubenswrapper[4818]: I1122 05:04:51.264466 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:04:51 crc kubenswrapper[4818]: I1122 05:04:51.264560 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:04:51 crc kubenswrapper[4818]: I1122 05:04:51.264633 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 05:04:51 crc kubenswrapper[4818]: I1122 05:04:51.265741 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"17a029baa87e074670137f3b22e9f934cbfb0d28077e00586b8813874c289d90"} pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 05:04:51 crc kubenswrapper[4818]: I1122 05:04:51.265821 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" containerID="cri-o://17a029baa87e074670137f3b22e9f934cbfb0d28077e00586b8813874c289d90" gracePeriod=600 Nov 22 05:04:51 crc kubenswrapper[4818]: I1122 05:04:51.649760 4818 generic.go:334] "Generic (PLEG): container finished" podID="0971c9ec-29ce-44eb-9875-ad4664143305" containerID="17a029baa87e074670137f3b22e9f934cbfb0d28077e00586b8813874c289d90" exitCode=0 Nov 22 05:04:51 crc kubenswrapper[4818]: I1122 05:04:51.649841 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerDied","Data":"17a029baa87e074670137f3b22e9f934cbfb0d28077e00586b8813874c289d90"} Nov 22 05:04:51 crc kubenswrapper[4818]: I1122 05:04:51.649899 4818 scope.go:117] "RemoveContainer" containerID="0c9825f8f0cd7e721d11fe80d62d92bcabce6091d17bb472097983d66a29cf6a" Nov 22 05:04:54 crc kubenswrapper[4818]: I1122 05:04:54.667820 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"7f4e621532f88c0da5dac82557bd8694f552742e853952ce47999bc11fbbc049"} Nov 22 05:04:54 crc kubenswrapper[4818]: I1122 05:04:54.671811 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5b7cbb7cc5-ff54p" event={"ID":"254a82ef-b456-4309-93fb-a236494bf3d4","Type":"ContainerStarted","Data":"f351630013476caec31b20fda206f431beb79239b0c96044c9f5d8a8e886f72c"} Nov 22 05:04:54 crc kubenswrapper[4818]: I1122 05:04:54.671895 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5b7cbb7cc5-ff54p" Nov 22 05:04:54 crc kubenswrapper[4818]: I1122 05:04:54.675433 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-757d8876f9-6qjkv" event={"ID":"63241779-80e0-43a3-bb9c-6530a4a6ae05","Type":"ContainerStarted","Data":"048ea79adb3ae8be708a291b729d2f8d1dba89b8e675c85a776dd8be348f8ec1"} Nov 22 05:04:54 crc kubenswrapper[4818]: I1122 05:04:54.675608 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-757d8876f9-6qjkv" Nov 22 05:04:54 crc kubenswrapper[4818]: I1122 05:04:54.700222 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-757d8876f9-6qjkv" podStartSLOduration=1.666325069 podStartE2EDuration="7.700205387s" podCreationTimestamp="2025-11-22 05:04:47 +0000 UTC" firstStartedPulling="2025-11-22 05:04:48.081713786 +0000 UTC m=+1040.656130313" lastFinishedPulling="2025-11-22 05:04:54.115594104 +0000 UTC m=+1046.690010631" observedRunningTime="2025-11-22 05:04:54.698540382 +0000 UTC m=+1047.272956919" watchObservedRunningTime="2025-11-22 05:04:54.700205387 +0000 UTC m=+1047.274621914" Nov 22 05:04:54 crc kubenswrapper[4818]: I1122 05:04:54.723636 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5b7cbb7cc5-ff54p" podStartSLOduration=1.504446275 podStartE2EDuration="7.723615229s" podCreationTimestamp="2025-11-22 05:04:47 +0000 UTC" firstStartedPulling="2025-11-22 05:04:47.877680442 +0000 UTC m=+1040.452096969" lastFinishedPulling="2025-11-22 05:04:54.096849396 +0000 UTC m=+1046.671265923" observedRunningTime="2025-11-22 05:04:54.72024544 +0000 UTC m=+1047.294661967" watchObservedRunningTime="2025-11-22 05:04:54.723615229 +0000 UTC m=+1047.298031766" Nov 22 05:04:56 crc kubenswrapper[4818]: I1122 05:04:56.354642 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-94k9s" Nov 22 05:04:56 crc kubenswrapper[4818]: I1122 05:04:56.354991 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-94k9s" Nov 22 05:04:56 crc kubenswrapper[4818]: I1122 05:04:56.401548 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-94k9s" Nov 22 05:04:56 crc kubenswrapper[4818]: I1122 05:04:56.723688 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-94k9s" Nov 22 05:04:56 crc kubenswrapper[4818]: I1122 05:04:56.766324 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-94k9s"] Nov 22 05:04:58 crc kubenswrapper[4818]: I1122 05:04:58.698272 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-94k9s" podUID="64fdc198-7025-4a93-903f-0a54a1f80967" containerName="registry-server" containerID="cri-o://6e8d8dcf70a4b8b8921705a1c4e6e8b4614f7c827d480b36c185afc49ab61fe4" gracePeriod=2 Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.131430 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-94k9s" Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.316298 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64fdc198-7025-4a93-903f-0a54a1f80967-utilities\") pod \"64fdc198-7025-4a93-903f-0a54a1f80967\" (UID: \"64fdc198-7025-4a93-903f-0a54a1f80967\") " Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.316407 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8zbv\" (UniqueName: \"kubernetes.io/projected/64fdc198-7025-4a93-903f-0a54a1f80967-kube-api-access-k8zbv\") pod \"64fdc198-7025-4a93-903f-0a54a1f80967\" (UID: \"64fdc198-7025-4a93-903f-0a54a1f80967\") " Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.316464 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64fdc198-7025-4a93-903f-0a54a1f80967-catalog-content\") pod \"64fdc198-7025-4a93-903f-0a54a1f80967\" (UID: \"64fdc198-7025-4a93-903f-0a54a1f80967\") " Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.317289 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64fdc198-7025-4a93-903f-0a54a1f80967-utilities" (OuterVolumeSpecName: "utilities") pod "64fdc198-7025-4a93-903f-0a54a1f80967" (UID: "64fdc198-7025-4a93-903f-0a54a1f80967"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.322103 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64fdc198-7025-4a93-903f-0a54a1f80967-kube-api-access-k8zbv" (OuterVolumeSpecName: "kube-api-access-k8zbv") pod "64fdc198-7025-4a93-903f-0a54a1f80967" (UID: "64fdc198-7025-4a93-903f-0a54a1f80967"). InnerVolumeSpecName "kube-api-access-k8zbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.371856 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64fdc198-7025-4a93-903f-0a54a1f80967-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "64fdc198-7025-4a93-903f-0a54a1f80967" (UID: "64fdc198-7025-4a93-903f-0a54a1f80967"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.417880 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64fdc198-7025-4a93-903f-0a54a1f80967-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.418010 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64fdc198-7025-4a93-903f-0a54a1f80967-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.418125 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8zbv\" (UniqueName: \"kubernetes.io/projected/64fdc198-7025-4a93-903f-0a54a1f80967-kube-api-access-k8zbv\") on node \"crc\" DevicePath \"\"" Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.707781 4818 generic.go:334] "Generic (PLEG): container finished" podID="64fdc198-7025-4a93-903f-0a54a1f80967" containerID="6e8d8dcf70a4b8b8921705a1c4e6e8b4614f7c827d480b36c185afc49ab61fe4" exitCode=0 Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.707828 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-94k9s" event={"ID":"64fdc198-7025-4a93-903f-0a54a1f80967","Type":"ContainerDied","Data":"6e8d8dcf70a4b8b8921705a1c4e6e8b4614f7c827d480b36c185afc49ab61fe4"} Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.707856 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-94k9s" event={"ID":"64fdc198-7025-4a93-903f-0a54a1f80967","Type":"ContainerDied","Data":"a380a2972105f8755c98a45bd44044d39c18de76d676dee2b87c2006915e5927"} Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.707877 4818 scope.go:117] "RemoveContainer" containerID="6e8d8dcf70a4b8b8921705a1c4e6e8b4614f7c827d480b36c185afc49ab61fe4" Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.708000 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-94k9s" Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.729649 4818 scope.go:117] "RemoveContainer" containerID="b87d2e5f6d106f7819d5241723b71cb7418d1fd7508f9960afea370e7f14cd5e" Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.746534 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-94k9s"] Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.751276 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-94k9s"] Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.767739 4818 scope.go:117] "RemoveContainer" containerID="4755e538b235b383b48be6b85296cfd969bae074dbac2ea37928a22d8d3d551f" Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.793403 4818 scope.go:117] "RemoveContainer" containerID="6e8d8dcf70a4b8b8921705a1c4e6e8b4614f7c827d480b36c185afc49ab61fe4" Nov 22 05:04:59 crc kubenswrapper[4818]: E1122 05:04:59.793928 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e8d8dcf70a4b8b8921705a1c4e6e8b4614f7c827d480b36c185afc49ab61fe4\": container with ID starting with 6e8d8dcf70a4b8b8921705a1c4e6e8b4614f7c827d480b36c185afc49ab61fe4 not found: ID does not exist" containerID="6e8d8dcf70a4b8b8921705a1c4e6e8b4614f7c827d480b36c185afc49ab61fe4" Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.793965 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e8d8dcf70a4b8b8921705a1c4e6e8b4614f7c827d480b36c185afc49ab61fe4"} err="failed to get container status \"6e8d8dcf70a4b8b8921705a1c4e6e8b4614f7c827d480b36c185afc49ab61fe4\": rpc error: code = NotFound desc = could not find container \"6e8d8dcf70a4b8b8921705a1c4e6e8b4614f7c827d480b36c185afc49ab61fe4\": container with ID starting with 6e8d8dcf70a4b8b8921705a1c4e6e8b4614f7c827d480b36c185afc49ab61fe4 not found: ID does not exist" Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.793992 4818 scope.go:117] "RemoveContainer" containerID="b87d2e5f6d106f7819d5241723b71cb7418d1fd7508f9960afea370e7f14cd5e" Nov 22 05:04:59 crc kubenswrapper[4818]: E1122 05:04:59.794332 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b87d2e5f6d106f7819d5241723b71cb7418d1fd7508f9960afea370e7f14cd5e\": container with ID starting with b87d2e5f6d106f7819d5241723b71cb7418d1fd7508f9960afea370e7f14cd5e not found: ID does not exist" containerID="b87d2e5f6d106f7819d5241723b71cb7418d1fd7508f9960afea370e7f14cd5e" Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.794362 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b87d2e5f6d106f7819d5241723b71cb7418d1fd7508f9960afea370e7f14cd5e"} err="failed to get container status \"b87d2e5f6d106f7819d5241723b71cb7418d1fd7508f9960afea370e7f14cd5e\": rpc error: code = NotFound desc = could not find container \"b87d2e5f6d106f7819d5241723b71cb7418d1fd7508f9960afea370e7f14cd5e\": container with ID starting with b87d2e5f6d106f7819d5241723b71cb7418d1fd7508f9960afea370e7f14cd5e not found: ID does not exist" Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.794379 4818 scope.go:117] "RemoveContainer" containerID="4755e538b235b383b48be6b85296cfd969bae074dbac2ea37928a22d8d3d551f" Nov 22 05:04:59 crc kubenswrapper[4818]: E1122 05:04:59.794915 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4755e538b235b383b48be6b85296cfd969bae074dbac2ea37928a22d8d3d551f\": container with ID starting with 4755e538b235b383b48be6b85296cfd969bae074dbac2ea37928a22d8d3d551f not found: ID does not exist" containerID="4755e538b235b383b48be6b85296cfd969bae074dbac2ea37928a22d8d3d551f" Nov 22 05:04:59 crc kubenswrapper[4818]: I1122 05:04:59.794936 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4755e538b235b383b48be6b85296cfd969bae074dbac2ea37928a22d8d3d551f"} err="failed to get container status \"4755e538b235b383b48be6b85296cfd969bae074dbac2ea37928a22d8d3d551f\": rpc error: code = NotFound desc = could not find container \"4755e538b235b383b48be6b85296cfd969bae074dbac2ea37928a22d8d3d551f\": container with ID starting with 4755e538b235b383b48be6b85296cfd969bae074dbac2ea37928a22d8d3d551f not found: ID does not exist" Nov 22 05:05:00 crc kubenswrapper[4818]: I1122 05:05:00.300327 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64fdc198-7025-4a93-903f-0a54a1f80967" path="/var/lib/kubelet/pods/64fdc198-7025-4a93-903f-0a54a1f80967/volumes" Nov 22 05:05:07 crc kubenswrapper[4818]: I1122 05:05:07.899343 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-757d8876f9-6qjkv" Nov 22 05:05:27 crc kubenswrapper[4818]: I1122 05:05:27.625964 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5b7cbb7cc5-ff54p" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.489670 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-ld8qn"] Nov 22 05:05:28 crc kubenswrapper[4818]: E1122 05:05:28.490143 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64fdc198-7025-4a93-903f-0a54a1f80967" containerName="extract-utilities" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.490159 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="64fdc198-7025-4a93-903f-0a54a1f80967" containerName="extract-utilities" Nov 22 05:05:28 crc kubenswrapper[4818]: E1122 05:05:28.490171 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64fdc198-7025-4a93-903f-0a54a1f80967" containerName="extract-content" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.490178 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="64fdc198-7025-4a93-903f-0a54a1f80967" containerName="extract-content" Nov 22 05:05:28 crc kubenswrapper[4818]: E1122 05:05:28.490183 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64fdc198-7025-4a93-903f-0a54a1f80967" containerName="registry-server" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.490189 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="64fdc198-7025-4a93-903f-0a54a1f80967" containerName="registry-server" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.490322 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="64fdc198-7025-4a93-903f-0a54a1f80967" containerName="registry-server" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.492267 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.494174 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-rss4n"] Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.494832 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.494962 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.495101 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-rss4n" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.495560 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-5b9fb" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.499666 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.504523 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-rss4n"] Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.540608 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svrbc\" (UniqueName: \"kubernetes.io/projected/a90765d9-f4ec-4f76-921b-f8fe52fa2cdc-kube-api-access-svrbc\") pod \"frr-k8s-webhook-server-6998585d5-rss4n\" (UID: \"a90765d9-f4ec-4f76-921b-f8fe52fa2cdc\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-rss4n" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.540654 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4bf0b266-290e-4c4f-8ddb-17c38155751e-frr-sockets\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.540685 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4bf0b266-290e-4c4f-8ddb-17c38155751e-metrics\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.540762 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4bf0b266-290e-4c4f-8ddb-17c38155751e-frr-startup\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.540868 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4bf0b266-290e-4c4f-8ddb-17c38155751e-frr-conf\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.540903 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxbjw\" (UniqueName: \"kubernetes.io/projected/4bf0b266-290e-4c4f-8ddb-17c38155751e-kube-api-access-hxbjw\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.540921 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4bf0b266-290e-4c4f-8ddb-17c38155751e-metrics-certs\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.540960 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4bf0b266-290e-4c4f-8ddb-17c38155751e-reloader\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.541142 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a90765d9-f4ec-4f76-921b-f8fe52fa2cdc-cert\") pod \"frr-k8s-webhook-server-6998585d5-rss4n\" (UID: \"a90765d9-f4ec-4f76-921b-f8fe52fa2cdc\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-rss4n" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.566537 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-8mwxx"] Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.567627 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-8mwxx" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.571321 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-sv558" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.571499 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.573417 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.574521 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.586533 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-wbzs6"] Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.587337 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-wbzs6" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.596377 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.606840 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-wbzs6"] Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.663636 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a90765d9-f4ec-4f76-921b-f8fe52fa2cdc-cert\") pod \"frr-k8s-webhook-server-6998585d5-rss4n\" (UID: \"a90765d9-f4ec-4f76-921b-f8fe52fa2cdc\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-rss4n" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.663705 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svrbc\" (UniqueName: \"kubernetes.io/projected/a90765d9-f4ec-4f76-921b-f8fe52fa2cdc-kube-api-access-svrbc\") pod \"frr-k8s-webhook-server-6998585d5-rss4n\" (UID: \"a90765d9-f4ec-4f76-921b-f8fe52fa2cdc\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-rss4n" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.663738 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4bf0b266-290e-4c4f-8ddb-17c38155751e-frr-sockets\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.664153 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4bf0b266-290e-4c4f-8ddb-17c38155751e-metrics\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.664213 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4bf0b266-290e-4c4f-8ddb-17c38155751e-frr-startup\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.664223 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4bf0b266-290e-4c4f-8ddb-17c38155751e-frr-sockets\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.664322 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4bf0b266-290e-4c4f-8ddb-17c38155751e-frr-conf\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.664367 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxbjw\" (UniqueName: \"kubernetes.io/projected/4bf0b266-290e-4c4f-8ddb-17c38155751e-kube-api-access-hxbjw\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.664391 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4bf0b266-290e-4c4f-8ddb-17c38155751e-metrics-certs\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.664442 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4bf0b266-290e-4c4f-8ddb-17c38155751e-reloader\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.664483 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4bf0b266-290e-4c4f-8ddb-17c38155751e-metrics\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: E1122 05:05:28.664619 4818 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Nov 22 05:05:28 crc kubenswrapper[4818]: E1122 05:05:28.664673 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bf0b266-290e-4c4f-8ddb-17c38155751e-metrics-certs podName:4bf0b266-290e-4c4f-8ddb-17c38155751e nodeName:}" failed. No retries permitted until 2025-11-22 05:05:29.164656706 +0000 UTC m=+1081.739073233 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4bf0b266-290e-4c4f-8ddb-17c38155751e-metrics-certs") pod "frr-k8s-ld8qn" (UID: "4bf0b266-290e-4c4f-8ddb-17c38155751e") : secret "frr-k8s-certs-secret" not found Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.664751 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4bf0b266-290e-4c4f-8ddb-17c38155751e-frr-conf\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.664767 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4bf0b266-290e-4c4f-8ddb-17c38155751e-reloader\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.664973 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4bf0b266-290e-4c4f-8ddb-17c38155751e-frr-startup\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.670695 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a90765d9-f4ec-4f76-921b-f8fe52fa2cdc-cert\") pod \"frr-k8s-webhook-server-6998585d5-rss4n\" (UID: \"a90765d9-f4ec-4f76-921b-f8fe52fa2cdc\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-rss4n" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.688003 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxbjw\" (UniqueName: \"kubernetes.io/projected/4bf0b266-290e-4c4f-8ddb-17c38155751e-kube-api-access-hxbjw\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.688897 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svrbc\" (UniqueName: \"kubernetes.io/projected/a90765d9-f4ec-4f76-921b-f8fe52fa2cdc-kube-api-access-svrbc\") pod \"frr-k8s-webhook-server-6998585d5-rss4n\" (UID: \"a90765d9-f4ec-4f76-921b-f8fe52fa2cdc\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-rss4n" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.765496 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/dbc8ab54-ce5e-4320-8deb-4d597cb4f82a-metallb-excludel2\") pod \"speaker-8mwxx\" (UID: \"dbc8ab54-ce5e-4320-8deb-4d597cb4f82a\") " pod="metallb-system/speaker-8mwxx" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.765715 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dbc8ab54-ce5e-4320-8deb-4d597cb4f82a-metrics-certs\") pod \"speaker-8mwxx\" (UID: \"dbc8ab54-ce5e-4320-8deb-4d597cb4f82a\") " pod="metallb-system/speaker-8mwxx" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.765765 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhlql\" (UniqueName: \"kubernetes.io/projected/dbc8ab54-ce5e-4320-8deb-4d597cb4f82a-kube-api-access-nhlql\") pod \"speaker-8mwxx\" (UID: \"dbc8ab54-ce5e-4320-8deb-4d597cb4f82a\") " pod="metallb-system/speaker-8mwxx" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.765838 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0fdd1082-4473-4604-8ced-2294bf4ba248-cert\") pod \"controller-6c7b4b5f48-wbzs6\" (UID: \"0fdd1082-4473-4604-8ced-2294bf4ba248\") " pod="metallb-system/controller-6c7b4b5f48-wbzs6" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.765945 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/dbc8ab54-ce5e-4320-8deb-4d597cb4f82a-memberlist\") pod \"speaker-8mwxx\" (UID: \"dbc8ab54-ce5e-4320-8deb-4d597cb4f82a\") " pod="metallb-system/speaker-8mwxx" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.765981 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0fdd1082-4473-4604-8ced-2294bf4ba248-metrics-certs\") pod \"controller-6c7b4b5f48-wbzs6\" (UID: \"0fdd1082-4473-4604-8ced-2294bf4ba248\") " pod="metallb-system/controller-6c7b4b5f48-wbzs6" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.766019 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l95z7\" (UniqueName: \"kubernetes.io/projected/0fdd1082-4473-4604-8ced-2294bf4ba248-kube-api-access-l95z7\") pod \"controller-6c7b4b5f48-wbzs6\" (UID: \"0fdd1082-4473-4604-8ced-2294bf4ba248\") " pod="metallb-system/controller-6c7b4b5f48-wbzs6" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.825116 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-rss4n" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.866552 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0fdd1082-4473-4604-8ced-2294bf4ba248-metrics-certs\") pod \"controller-6c7b4b5f48-wbzs6\" (UID: \"0fdd1082-4473-4604-8ced-2294bf4ba248\") " pod="metallb-system/controller-6c7b4b5f48-wbzs6" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.866600 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l95z7\" (UniqueName: \"kubernetes.io/projected/0fdd1082-4473-4604-8ced-2294bf4ba248-kube-api-access-l95z7\") pod \"controller-6c7b4b5f48-wbzs6\" (UID: \"0fdd1082-4473-4604-8ced-2294bf4ba248\") " pod="metallb-system/controller-6c7b4b5f48-wbzs6" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.866624 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/dbc8ab54-ce5e-4320-8deb-4d597cb4f82a-metallb-excludel2\") pod \"speaker-8mwxx\" (UID: \"dbc8ab54-ce5e-4320-8deb-4d597cb4f82a\") " pod="metallb-system/speaker-8mwxx" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.866652 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dbc8ab54-ce5e-4320-8deb-4d597cb4f82a-metrics-certs\") pod \"speaker-8mwxx\" (UID: \"dbc8ab54-ce5e-4320-8deb-4d597cb4f82a\") " pod="metallb-system/speaker-8mwxx" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.866666 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhlql\" (UniqueName: \"kubernetes.io/projected/dbc8ab54-ce5e-4320-8deb-4d597cb4f82a-kube-api-access-nhlql\") pod \"speaker-8mwxx\" (UID: \"dbc8ab54-ce5e-4320-8deb-4d597cb4f82a\") " pod="metallb-system/speaker-8mwxx" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.866691 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0fdd1082-4473-4604-8ced-2294bf4ba248-cert\") pod \"controller-6c7b4b5f48-wbzs6\" (UID: \"0fdd1082-4473-4604-8ced-2294bf4ba248\") " pod="metallb-system/controller-6c7b4b5f48-wbzs6" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.866731 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/dbc8ab54-ce5e-4320-8deb-4d597cb4f82a-memberlist\") pod \"speaker-8mwxx\" (UID: \"dbc8ab54-ce5e-4320-8deb-4d597cb4f82a\") " pod="metallb-system/speaker-8mwxx" Nov 22 05:05:28 crc kubenswrapper[4818]: E1122 05:05:28.866824 4818 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 22 05:05:28 crc kubenswrapper[4818]: E1122 05:05:28.866866 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dbc8ab54-ce5e-4320-8deb-4d597cb4f82a-memberlist podName:dbc8ab54-ce5e-4320-8deb-4d597cb4f82a nodeName:}" failed. No retries permitted until 2025-11-22 05:05:29.366853132 +0000 UTC m=+1081.941269659 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/dbc8ab54-ce5e-4320-8deb-4d597cb4f82a-memberlist") pod "speaker-8mwxx" (UID: "dbc8ab54-ce5e-4320-8deb-4d597cb4f82a") : secret "metallb-memberlist" not found Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.868462 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/dbc8ab54-ce5e-4320-8deb-4d597cb4f82a-metallb-excludel2\") pod \"speaker-8mwxx\" (UID: \"dbc8ab54-ce5e-4320-8deb-4d597cb4f82a\") " pod="metallb-system/speaker-8mwxx" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.868942 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.872561 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dbc8ab54-ce5e-4320-8deb-4d597cb4f82a-metrics-certs\") pod \"speaker-8mwxx\" (UID: \"dbc8ab54-ce5e-4320-8deb-4d597cb4f82a\") " pod="metallb-system/speaker-8mwxx" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.873379 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0fdd1082-4473-4604-8ced-2294bf4ba248-metrics-certs\") pod \"controller-6c7b4b5f48-wbzs6\" (UID: \"0fdd1082-4473-4604-8ced-2294bf4ba248\") " pod="metallb-system/controller-6c7b4b5f48-wbzs6" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.880157 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0fdd1082-4473-4604-8ced-2294bf4ba248-cert\") pod \"controller-6c7b4b5f48-wbzs6\" (UID: \"0fdd1082-4473-4604-8ced-2294bf4ba248\") " pod="metallb-system/controller-6c7b4b5f48-wbzs6" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.882860 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhlql\" (UniqueName: \"kubernetes.io/projected/dbc8ab54-ce5e-4320-8deb-4d597cb4f82a-kube-api-access-nhlql\") pod \"speaker-8mwxx\" (UID: \"dbc8ab54-ce5e-4320-8deb-4d597cb4f82a\") " pod="metallb-system/speaker-8mwxx" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.884926 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l95z7\" (UniqueName: \"kubernetes.io/projected/0fdd1082-4473-4604-8ced-2294bf4ba248-kube-api-access-l95z7\") pod \"controller-6c7b4b5f48-wbzs6\" (UID: \"0fdd1082-4473-4604-8ced-2294bf4ba248\") " pod="metallb-system/controller-6c7b4b5f48-wbzs6" Nov 22 05:05:28 crc kubenswrapper[4818]: I1122 05:05:28.898816 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-wbzs6" Nov 22 05:05:29 crc kubenswrapper[4818]: I1122 05:05:29.049679 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-rss4n"] Nov 22 05:05:29 crc kubenswrapper[4818]: I1122 05:05:29.169558 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4bf0b266-290e-4c4f-8ddb-17c38155751e-metrics-certs\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:29 crc kubenswrapper[4818]: I1122 05:05:29.174449 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4bf0b266-290e-4c4f-8ddb-17c38155751e-metrics-certs\") pod \"frr-k8s-ld8qn\" (UID: \"4bf0b266-290e-4c4f-8ddb-17c38155751e\") " pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:29 crc kubenswrapper[4818]: I1122 05:05:29.347538 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-wbzs6"] Nov 22 05:05:29 crc kubenswrapper[4818]: W1122 05:05:29.348169 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0fdd1082_4473_4604_8ced_2294bf4ba248.slice/crio-55addbf3e5bbbb4208edc276db4ff749785be605bd35109819d6c4526c17879b WatchSource:0}: Error finding container 55addbf3e5bbbb4208edc276db4ff749785be605bd35109819d6c4526c17879b: Status 404 returned error can't find the container with id 55addbf3e5bbbb4208edc276db4ff749785be605bd35109819d6c4526c17879b Nov 22 05:05:29 crc kubenswrapper[4818]: I1122 05:05:29.372503 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/dbc8ab54-ce5e-4320-8deb-4d597cb4f82a-memberlist\") pod \"speaker-8mwxx\" (UID: \"dbc8ab54-ce5e-4320-8deb-4d597cb4f82a\") " pod="metallb-system/speaker-8mwxx" Nov 22 05:05:29 crc kubenswrapper[4818]: E1122 05:05:29.372730 4818 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 22 05:05:29 crc kubenswrapper[4818]: E1122 05:05:29.372914 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dbc8ab54-ce5e-4320-8deb-4d597cb4f82a-memberlist podName:dbc8ab54-ce5e-4320-8deb-4d597cb4f82a nodeName:}" failed. No retries permitted until 2025-11-22 05:05:30.372814803 +0000 UTC m=+1082.947231340 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/dbc8ab54-ce5e-4320-8deb-4d597cb4f82a-memberlist") pod "speaker-8mwxx" (UID: "dbc8ab54-ce5e-4320-8deb-4d597cb4f82a") : secret "metallb-memberlist" not found Nov 22 05:05:29 crc kubenswrapper[4818]: I1122 05:05:29.419562 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:29 crc kubenswrapper[4818]: I1122 05:05:29.907960 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-wbzs6" event={"ID":"0fdd1082-4473-4604-8ced-2294bf4ba248","Type":"ContainerStarted","Data":"f81ca86a2216b37007b297e665560af68f917424afc6eeb215ae9c5fe258aa5a"} Nov 22 05:05:29 crc kubenswrapper[4818]: I1122 05:05:29.908531 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-wbzs6" event={"ID":"0fdd1082-4473-4604-8ced-2294bf4ba248","Type":"ContainerStarted","Data":"55addbf3e5bbbb4208edc276db4ff749785be605bd35109819d6c4526c17879b"} Nov 22 05:05:29 crc kubenswrapper[4818]: I1122 05:05:29.909437 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-rss4n" event={"ID":"a90765d9-f4ec-4f76-921b-f8fe52fa2cdc","Type":"ContainerStarted","Data":"217199bf18041fbf73d32b85fe55243fdfa6ff595e1be9b138204b98cbf280df"} Nov 22 05:05:29 crc kubenswrapper[4818]: I1122 05:05:29.910326 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ld8qn" event={"ID":"4bf0b266-290e-4c4f-8ddb-17c38155751e","Type":"ContainerStarted","Data":"896288b61b74e8009d79931d982f18aa61ae2e145ce782636f1decc29f05ee01"} Nov 22 05:05:30 crc kubenswrapper[4818]: I1122 05:05:30.385899 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/dbc8ab54-ce5e-4320-8deb-4d597cb4f82a-memberlist\") pod \"speaker-8mwxx\" (UID: \"dbc8ab54-ce5e-4320-8deb-4d597cb4f82a\") " pod="metallb-system/speaker-8mwxx" Nov 22 05:05:30 crc kubenswrapper[4818]: I1122 05:05:30.396283 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/dbc8ab54-ce5e-4320-8deb-4d597cb4f82a-memberlist\") pod \"speaker-8mwxx\" (UID: \"dbc8ab54-ce5e-4320-8deb-4d597cb4f82a\") " pod="metallb-system/speaker-8mwxx" Nov 22 05:05:30 crc kubenswrapper[4818]: I1122 05:05:30.681038 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-8mwxx" Nov 22 05:05:30 crc kubenswrapper[4818]: I1122 05:05:30.917724 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-wbzs6" event={"ID":"0fdd1082-4473-4604-8ced-2294bf4ba248","Type":"ContainerStarted","Data":"e6cd4f5f5de5e27fe7fdf47bcd8e2a2bfeffff7e1b54be360e5ea68d631a4582"} Nov 22 05:05:30 crc kubenswrapper[4818]: I1122 05:05:30.917809 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-wbzs6" Nov 22 05:05:30 crc kubenswrapper[4818]: I1122 05:05:30.920839 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8mwxx" event={"ID":"dbc8ab54-ce5e-4320-8deb-4d597cb4f82a","Type":"ContainerStarted","Data":"91da3293a76684e0b3b54bdddec49462c1916917dbfa5cbb97d659402caf1154"} Nov 22 05:05:30 crc kubenswrapper[4818]: I1122 05:05:30.944946 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-wbzs6" podStartSLOduration=2.94492608 podStartE2EDuration="2.94492608s" podCreationTimestamp="2025-11-22 05:05:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:05:30.935241753 +0000 UTC m=+1083.509658300" watchObservedRunningTime="2025-11-22 05:05:30.94492608 +0000 UTC m=+1083.519342607" Nov 22 05:05:31 crc kubenswrapper[4818]: I1122 05:05:31.931480 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8mwxx" event={"ID":"dbc8ab54-ce5e-4320-8deb-4d597cb4f82a","Type":"ContainerStarted","Data":"78843a7aa6b4e292e481c1108382f6d146de85013ab43a9ae18347fdc83635ba"} Nov 22 05:05:32 crc kubenswrapper[4818]: I1122 05:05:32.937692 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8mwxx" event={"ID":"dbc8ab54-ce5e-4320-8deb-4d597cb4f82a","Type":"ContainerStarted","Data":"2dc5a14870128cfb652dec65c9836b2e596b6b7a63b865ab926e4de5afe97234"} Nov 22 05:05:32 crc kubenswrapper[4818]: I1122 05:05:32.939315 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-8mwxx" Nov 22 05:05:32 crc kubenswrapper[4818]: I1122 05:05:32.961586 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-8mwxx" podStartSLOduration=4.961567885 podStartE2EDuration="4.961567885s" podCreationTimestamp="2025-11-22 05:05:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:05:32.959609583 +0000 UTC m=+1085.534026110" watchObservedRunningTime="2025-11-22 05:05:32.961567885 +0000 UTC m=+1085.535984412" Nov 22 05:05:40 crc kubenswrapper[4818]: I1122 05:05:40.995809 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-rss4n" event={"ID":"a90765d9-f4ec-4f76-921b-f8fe52fa2cdc","Type":"ContainerStarted","Data":"85d56084ee05ea89a8c4c861cbc5ea9e7adeb2db731a15aa45e49e9d4c56db72"} Nov 22 05:05:40 crc kubenswrapper[4818]: I1122 05:05:40.996408 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-rss4n" Nov 22 05:05:40 crc kubenswrapper[4818]: I1122 05:05:40.997643 4818 generic.go:334] "Generic (PLEG): container finished" podID="4bf0b266-290e-4c4f-8ddb-17c38155751e" containerID="d6fd1df77d8daaf41c30fecbb30d734ef859e06b5551e4928bc7f6ef5ce6e00a" exitCode=0 Nov 22 05:05:40 crc kubenswrapper[4818]: I1122 05:05:40.997687 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ld8qn" event={"ID":"4bf0b266-290e-4c4f-8ddb-17c38155751e","Type":"ContainerDied","Data":"d6fd1df77d8daaf41c30fecbb30d734ef859e06b5551e4928bc7f6ef5ce6e00a"} Nov 22 05:05:41 crc kubenswrapper[4818]: I1122 05:05:41.013204 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-rss4n" podStartSLOduration=2.459039752 podStartE2EDuration="13.013187028s" podCreationTimestamp="2025-11-22 05:05:28 +0000 UTC" firstStartedPulling="2025-11-22 05:05:29.066462188 +0000 UTC m=+1081.640878715" lastFinishedPulling="2025-11-22 05:05:39.620609424 +0000 UTC m=+1092.195025991" observedRunningTime="2025-11-22 05:05:41.010781794 +0000 UTC m=+1093.585198331" watchObservedRunningTime="2025-11-22 05:05:41.013187028 +0000 UTC m=+1093.587603555" Nov 22 05:05:42 crc kubenswrapper[4818]: I1122 05:05:42.005627 4818 generic.go:334] "Generic (PLEG): container finished" podID="4bf0b266-290e-4c4f-8ddb-17c38155751e" containerID="242fa7017f3219dbc7c22175dad43dc8c44051ce901cb4a2fc6e4d61fe171d91" exitCode=0 Nov 22 05:05:42 crc kubenswrapper[4818]: I1122 05:05:42.005766 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ld8qn" event={"ID":"4bf0b266-290e-4c4f-8ddb-17c38155751e","Type":"ContainerDied","Data":"242fa7017f3219dbc7c22175dad43dc8c44051ce901cb4a2fc6e4d61fe171d91"} Nov 22 05:05:43 crc kubenswrapper[4818]: I1122 05:05:43.015123 4818 generic.go:334] "Generic (PLEG): container finished" podID="4bf0b266-290e-4c4f-8ddb-17c38155751e" containerID="7f0b89ff4365ff4f44f9439ac7c54dda524aa872134419c68a8545e5cbfcad23" exitCode=0 Nov 22 05:05:43 crc kubenswrapper[4818]: I1122 05:05:43.015188 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ld8qn" event={"ID":"4bf0b266-290e-4c4f-8ddb-17c38155751e","Type":"ContainerDied","Data":"7f0b89ff4365ff4f44f9439ac7c54dda524aa872134419c68a8545e5cbfcad23"} Nov 22 05:05:44 crc kubenswrapper[4818]: I1122 05:05:44.023584 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ld8qn" event={"ID":"4bf0b266-290e-4c4f-8ddb-17c38155751e","Type":"ContainerStarted","Data":"3fd66e28d8d4136b8039ee45d15b81a33ae114013c209f07390251657df362e8"} Nov 22 05:05:44 crc kubenswrapper[4818]: I1122 05:05:44.023969 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ld8qn" event={"ID":"4bf0b266-290e-4c4f-8ddb-17c38155751e","Type":"ContainerStarted","Data":"5edc911d673bc5c2238b924d5b1c833422ae9bceff6037945a5cc2322c841fb1"} Nov 22 05:05:45 crc kubenswrapper[4818]: I1122 05:05:45.034857 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ld8qn" event={"ID":"4bf0b266-290e-4c4f-8ddb-17c38155751e","Type":"ContainerStarted","Data":"9dd18c4ffe0e9e48ddab7c0388e88a0715b318e59ce1f5f556539cd5ab35be8d"} Nov 22 05:05:45 crc kubenswrapper[4818]: I1122 05:05:45.034954 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ld8qn" event={"ID":"4bf0b266-290e-4c4f-8ddb-17c38155751e","Type":"ContainerStarted","Data":"12d56c29867ca60bd51fa9b32f9bd798c5de438b0486b35e0e5ccdff1758b0fc"} Nov 22 05:05:46 crc kubenswrapper[4818]: I1122 05:05:46.059554 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ld8qn" event={"ID":"4bf0b266-290e-4c4f-8ddb-17c38155751e","Type":"ContainerStarted","Data":"0c19cf66a402147efe8521b8221555f575d032bb055efc277a05de0479f1ca5f"} Nov 22 05:05:46 crc kubenswrapper[4818]: I1122 05:05:46.061607 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ld8qn" event={"ID":"4bf0b266-290e-4c4f-8ddb-17c38155751e","Type":"ContainerStarted","Data":"bf2355008e05f240f83d703e47e7ac586e44f08d5845edf3d8369ae7c03d8efd"} Nov 22 05:05:46 crc kubenswrapper[4818]: I1122 05:05:46.061674 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:46 crc kubenswrapper[4818]: I1122 05:05:46.096898 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-ld8qn" podStartSLOduration=8.305611692 podStartE2EDuration="18.096871445s" podCreationTimestamp="2025-11-22 05:05:28 +0000 UTC" firstStartedPulling="2025-11-22 05:05:29.861512396 +0000 UTC m=+1082.435928923" lastFinishedPulling="2025-11-22 05:05:39.652772109 +0000 UTC m=+1092.227188676" observedRunningTime="2025-11-22 05:05:46.096164705 +0000 UTC m=+1098.670581272" watchObservedRunningTime="2025-11-22 05:05:46.096871445 +0000 UTC m=+1098.671288012" Nov 22 05:05:48 crc kubenswrapper[4818]: I1122 05:05:48.906000 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-wbzs6" Nov 22 05:05:49 crc kubenswrapper[4818]: I1122 05:05:49.419884 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:49 crc kubenswrapper[4818]: I1122 05:05:49.456982 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:05:50 crc kubenswrapper[4818]: I1122 05:05:50.685051 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-8mwxx" Nov 22 05:05:53 crc kubenswrapper[4818]: I1122 05:05:53.684849 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-8g4zx"] Nov 22 05:05:53 crc kubenswrapper[4818]: I1122 05:05:53.687750 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-8g4zx" Nov 22 05:05:53 crc kubenswrapper[4818]: I1122 05:05:53.690775 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-xtq5c" Nov 22 05:05:53 crc kubenswrapper[4818]: I1122 05:05:53.691165 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 22 05:05:53 crc kubenswrapper[4818]: I1122 05:05:53.692575 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 22 05:05:53 crc kubenswrapper[4818]: I1122 05:05:53.720613 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-8g4zx"] Nov 22 05:05:53 crc kubenswrapper[4818]: I1122 05:05:53.734627 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkj67\" (UniqueName: \"kubernetes.io/projected/15e18857-bc54-438b-9551-fdb5146918f3-kube-api-access-dkj67\") pod \"openstack-operator-index-8g4zx\" (UID: \"15e18857-bc54-438b-9551-fdb5146918f3\") " pod="openstack-operators/openstack-operator-index-8g4zx" Nov 22 05:05:53 crc kubenswrapper[4818]: I1122 05:05:53.836004 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkj67\" (UniqueName: \"kubernetes.io/projected/15e18857-bc54-438b-9551-fdb5146918f3-kube-api-access-dkj67\") pod \"openstack-operator-index-8g4zx\" (UID: \"15e18857-bc54-438b-9551-fdb5146918f3\") " pod="openstack-operators/openstack-operator-index-8g4zx" Nov 22 05:05:53 crc kubenswrapper[4818]: I1122 05:05:53.857229 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkj67\" (UniqueName: \"kubernetes.io/projected/15e18857-bc54-438b-9551-fdb5146918f3-kube-api-access-dkj67\") pod \"openstack-operator-index-8g4zx\" (UID: \"15e18857-bc54-438b-9551-fdb5146918f3\") " pod="openstack-operators/openstack-operator-index-8g4zx" Nov 22 05:05:54 crc kubenswrapper[4818]: I1122 05:05:54.009707 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-8g4zx" Nov 22 05:05:54 crc kubenswrapper[4818]: I1122 05:05:54.446421 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-8g4zx"] Nov 22 05:05:55 crc kubenswrapper[4818]: I1122 05:05:55.129371 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-8g4zx" event={"ID":"15e18857-bc54-438b-9551-fdb5146918f3","Type":"ContainerStarted","Data":"a26c837d81a9e3ffd2f767cfdcee10e2dbcee58831bc1908a5a04a15f848e18d"} Nov 22 05:05:56 crc kubenswrapper[4818]: I1122 05:05:56.847293 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-8g4zx"] Nov 22 05:05:57 crc kubenswrapper[4818]: I1122 05:05:57.454150 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-tn592"] Nov 22 05:05:57 crc kubenswrapper[4818]: I1122 05:05:57.455233 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tn592" Nov 22 05:05:57 crc kubenswrapper[4818]: I1122 05:05:57.467188 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tn592"] Nov 22 05:05:57 crc kubenswrapper[4818]: I1122 05:05:57.590506 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfjx9\" (UniqueName: \"kubernetes.io/projected/124c736f-9e78-422d-b317-b7263a45c55e-kube-api-access-nfjx9\") pod \"openstack-operator-index-tn592\" (UID: \"124c736f-9e78-422d-b317-b7263a45c55e\") " pod="openstack-operators/openstack-operator-index-tn592" Nov 22 05:05:57 crc kubenswrapper[4818]: I1122 05:05:57.691462 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfjx9\" (UniqueName: \"kubernetes.io/projected/124c736f-9e78-422d-b317-b7263a45c55e-kube-api-access-nfjx9\") pod \"openstack-operator-index-tn592\" (UID: \"124c736f-9e78-422d-b317-b7263a45c55e\") " pod="openstack-operators/openstack-operator-index-tn592" Nov 22 05:05:57 crc kubenswrapper[4818]: I1122 05:05:57.714818 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfjx9\" (UniqueName: \"kubernetes.io/projected/124c736f-9e78-422d-b317-b7263a45c55e-kube-api-access-nfjx9\") pod \"openstack-operator-index-tn592\" (UID: \"124c736f-9e78-422d-b317-b7263a45c55e\") " pod="openstack-operators/openstack-operator-index-tn592" Nov 22 05:05:57 crc kubenswrapper[4818]: I1122 05:05:57.780519 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tn592" Nov 22 05:05:58 crc kubenswrapper[4818]: I1122 05:05:58.257316 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tn592"] Nov 22 05:05:58 crc kubenswrapper[4818]: W1122 05:05:58.263283 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod124c736f_9e78_422d_b317_b7263a45c55e.slice/crio-b74ec6961a7cfb17814edd14d5965656310944a0632b4df9273007600448d93e WatchSource:0}: Error finding container b74ec6961a7cfb17814edd14d5965656310944a0632b4df9273007600448d93e: Status 404 returned error can't find the container with id b74ec6961a7cfb17814edd14d5965656310944a0632b4df9273007600448d93e Nov 22 05:05:58 crc kubenswrapper[4818]: I1122 05:05:58.830126 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-rss4n" Nov 22 05:05:59 crc kubenswrapper[4818]: I1122 05:05:59.172718 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tn592" event={"ID":"124c736f-9e78-422d-b317-b7263a45c55e","Type":"ContainerStarted","Data":"b74ec6961a7cfb17814edd14d5965656310944a0632b4df9273007600448d93e"} Nov 22 05:05:59 crc kubenswrapper[4818]: I1122 05:05:59.425324 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-ld8qn" Nov 22 05:06:04 crc kubenswrapper[4818]: I1122 05:06:04.208814 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tn592" event={"ID":"124c736f-9e78-422d-b317-b7263a45c55e","Type":"ContainerStarted","Data":"257d23309b53b04c9ca0f05bedba3f0273fc249d1606f1160d9f28d0b6e411f2"} Nov 22 05:06:04 crc kubenswrapper[4818]: I1122 05:06:04.210951 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-8g4zx" event={"ID":"15e18857-bc54-438b-9551-fdb5146918f3","Type":"ContainerStarted","Data":"152f4a8b57bf780ff7de3cb92c60ec2b8f3f55f2c7dcee5b23b2681a95a1aa9d"} Nov 22 05:06:04 crc kubenswrapper[4818]: I1122 05:06:04.211128 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-8g4zx" podUID="15e18857-bc54-438b-9551-fdb5146918f3" containerName="registry-server" containerID="cri-o://152f4a8b57bf780ff7de3cb92c60ec2b8f3f55f2c7dcee5b23b2681a95a1aa9d" gracePeriod=2 Nov 22 05:06:04 crc kubenswrapper[4818]: I1122 05:06:04.277845 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-tn592" podStartSLOduration=2.226952055 podStartE2EDuration="7.277825379s" podCreationTimestamp="2025-11-22 05:05:57 +0000 UTC" firstStartedPulling="2025-11-22 05:05:58.266621793 +0000 UTC m=+1110.841038320" lastFinishedPulling="2025-11-22 05:06:03.317495117 +0000 UTC m=+1115.891911644" observedRunningTime="2025-11-22 05:06:04.262627265 +0000 UTC m=+1116.837043792" watchObservedRunningTime="2025-11-22 05:06:04.277825379 +0000 UTC m=+1116.852241906" Nov 22 05:06:04 crc kubenswrapper[4818]: I1122 05:06:04.278648 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-8g4zx" podStartSLOduration=2.5399023 podStartE2EDuration="11.27864071s" podCreationTimestamp="2025-11-22 05:05:53 +0000 UTC" firstStartedPulling="2025-11-22 05:05:54.451550735 +0000 UTC m=+1107.025967272" lastFinishedPulling="2025-11-22 05:06:03.190289155 +0000 UTC m=+1115.764705682" observedRunningTime="2025-11-22 05:06:04.275548488 +0000 UTC m=+1116.849965015" watchObservedRunningTime="2025-11-22 05:06:04.27864071 +0000 UTC m=+1116.853057237" Nov 22 05:06:04 crc kubenswrapper[4818]: I1122 05:06:04.633355 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-8g4zx" Nov 22 05:06:04 crc kubenswrapper[4818]: I1122 05:06:04.707010 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkj67\" (UniqueName: \"kubernetes.io/projected/15e18857-bc54-438b-9551-fdb5146918f3-kube-api-access-dkj67\") pod \"15e18857-bc54-438b-9551-fdb5146918f3\" (UID: \"15e18857-bc54-438b-9551-fdb5146918f3\") " Nov 22 05:06:04 crc kubenswrapper[4818]: I1122 05:06:04.712560 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15e18857-bc54-438b-9551-fdb5146918f3-kube-api-access-dkj67" (OuterVolumeSpecName: "kube-api-access-dkj67") pod "15e18857-bc54-438b-9551-fdb5146918f3" (UID: "15e18857-bc54-438b-9551-fdb5146918f3"). InnerVolumeSpecName "kube-api-access-dkj67". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:06:04 crc kubenswrapper[4818]: I1122 05:06:04.808024 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkj67\" (UniqueName: \"kubernetes.io/projected/15e18857-bc54-438b-9551-fdb5146918f3-kube-api-access-dkj67\") on node \"crc\" DevicePath \"\"" Nov 22 05:06:05 crc kubenswrapper[4818]: I1122 05:06:05.220438 4818 generic.go:334] "Generic (PLEG): container finished" podID="15e18857-bc54-438b-9551-fdb5146918f3" containerID="152f4a8b57bf780ff7de3cb92c60ec2b8f3f55f2c7dcee5b23b2681a95a1aa9d" exitCode=0 Nov 22 05:06:05 crc kubenswrapper[4818]: I1122 05:06:05.220499 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-8g4zx" event={"ID":"15e18857-bc54-438b-9551-fdb5146918f3","Type":"ContainerDied","Data":"152f4a8b57bf780ff7de3cb92c60ec2b8f3f55f2c7dcee5b23b2681a95a1aa9d"} Nov 22 05:06:05 crc kubenswrapper[4818]: I1122 05:06:05.220562 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-8g4zx" Nov 22 05:06:05 crc kubenswrapper[4818]: I1122 05:06:05.220580 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-8g4zx" event={"ID":"15e18857-bc54-438b-9551-fdb5146918f3","Type":"ContainerDied","Data":"a26c837d81a9e3ffd2f767cfdcee10e2dbcee58831bc1908a5a04a15f848e18d"} Nov 22 05:06:05 crc kubenswrapper[4818]: I1122 05:06:05.220612 4818 scope.go:117] "RemoveContainer" containerID="152f4a8b57bf780ff7de3cb92c60ec2b8f3f55f2c7dcee5b23b2681a95a1aa9d" Nov 22 05:06:05 crc kubenswrapper[4818]: I1122 05:06:05.240421 4818 scope.go:117] "RemoveContainer" containerID="152f4a8b57bf780ff7de3cb92c60ec2b8f3f55f2c7dcee5b23b2681a95a1aa9d" Nov 22 05:06:05 crc kubenswrapper[4818]: E1122 05:06:05.241055 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"152f4a8b57bf780ff7de3cb92c60ec2b8f3f55f2c7dcee5b23b2681a95a1aa9d\": container with ID starting with 152f4a8b57bf780ff7de3cb92c60ec2b8f3f55f2c7dcee5b23b2681a95a1aa9d not found: ID does not exist" containerID="152f4a8b57bf780ff7de3cb92c60ec2b8f3f55f2c7dcee5b23b2681a95a1aa9d" Nov 22 05:06:05 crc kubenswrapper[4818]: I1122 05:06:05.241117 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"152f4a8b57bf780ff7de3cb92c60ec2b8f3f55f2c7dcee5b23b2681a95a1aa9d"} err="failed to get container status \"152f4a8b57bf780ff7de3cb92c60ec2b8f3f55f2c7dcee5b23b2681a95a1aa9d\": rpc error: code = NotFound desc = could not find container \"152f4a8b57bf780ff7de3cb92c60ec2b8f3f55f2c7dcee5b23b2681a95a1aa9d\": container with ID starting with 152f4a8b57bf780ff7de3cb92c60ec2b8f3f55f2c7dcee5b23b2681a95a1aa9d not found: ID does not exist" Nov 22 05:06:05 crc kubenswrapper[4818]: I1122 05:06:05.262371 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-8g4zx"] Nov 22 05:06:05 crc kubenswrapper[4818]: I1122 05:06:05.267950 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-8g4zx"] Nov 22 05:06:06 crc kubenswrapper[4818]: I1122 05:06:06.300659 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15e18857-bc54-438b-9551-fdb5146918f3" path="/var/lib/kubelet/pods/15e18857-bc54-438b-9551-fdb5146918f3/volumes" Nov 22 05:06:07 crc kubenswrapper[4818]: I1122 05:06:07.781228 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-tn592" Nov 22 05:06:07 crc kubenswrapper[4818]: I1122 05:06:07.781346 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-tn592" Nov 22 05:06:07 crc kubenswrapper[4818]: I1122 05:06:07.809239 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-tn592" Nov 22 05:06:08 crc kubenswrapper[4818]: I1122 05:06:08.287309 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-tn592" Nov 22 05:06:14 crc kubenswrapper[4818]: I1122 05:06:14.420357 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw"] Nov 22 05:06:14 crc kubenswrapper[4818]: E1122 05:06:14.421010 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15e18857-bc54-438b-9551-fdb5146918f3" containerName="registry-server" Nov 22 05:06:14 crc kubenswrapper[4818]: I1122 05:06:14.421026 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="15e18857-bc54-438b-9551-fdb5146918f3" containerName="registry-server" Nov 22 05:06:14 crc kubenswrapper[4818]: I1122 05:06:14.421174 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="15e18857-bc54-438b-9551-fdb5146918f3" containerName="registry-server" Nov 22 05:06:14 crc kubenswrapper[4818]: I1122 05:06:14.422134 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw" Nov 22 05:06:14 crc kubenswrapper[4818]: I1122 05:06:14.424094 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-75nl7" Nov 22 05:06:14 crc kubenswrapper[4818]: I1122 05:06:14.433530 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw"] Nov 22 05:06:14 crc kubenswrapper[4818]: I1122 05:06:14.447426 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttrst\" (UniqueName: \"kubernetes.io/projected/a5e6bbd8-bd14-4785-addd-67c8028cbd22-kube-api-access-ttrst\") pod \"f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw\" (UID: \"a5e6bbd8-bd14-4785-addd-67c8028cbd22\") " pod="openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw" Nov 22 05:06:14 crc kubenswrapper[4818]: I1122 05:06:14.447499 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5e6bbd8-bd14-4785-addd-67c8028cbd22-util\") pod \"f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw\" (UID: \"a5e6bbd8-bd14-4785-addd-67c8028cbd22\") " pod="openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw" Nov 22 05:06:14 crc kubenswrapper[4818]: I1122 05:06:14.447542 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5e6bbd8-bd14-4785-addd-67c8028cbd22-bundle\") pod \"f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw\" (UID: \"a5e6bbd8-bd14-4785-addd-67c8028cbd22\") " pod="openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw" Nov 22 05:06:14 crc kubenswrapper[4818]: I1122 05:06:14.548804 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5e6bbd8-bd14-4785-addd-67c8028cbd22-util\") pod \"f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw\" (UID: \"a5e6bbd8-bd14-4785-addd-67c8028cbd22\") " pod="openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw" Nov 22 05:06:14 crc kubenswrapper[4818]: I1122 05:06:14.548871 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5e6bbd8-bd14-4785-addd-67c8028cbd22-bundle\") pod \"f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw\" (UID: \"a5e6bbd8-bd14-4785-addd-67c8028cbd22\") " pod="openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw" Nov 22 05:06:14 crc kubenswrapper[4818]: I1122 05:06:14.548937 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttrst\" (UniqueName: \"kubernetes.io/projected/a5e6bbd8-bd14-4785-addd-67c8028cbd22-kube-api-access-ttrst\") pod \"f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw\" (UID: \"a5e6bbd8-bd14-4785-addd-67c8028cbd22\") " pod="openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw" Nov 22 05:06:14 crc kubenswrapper[4818]: I1122 05:06:14.549682 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5e6bbd8-bd14-4785-addd-67c8028cbd22-util\") pod \"f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw\" (UID: \"a5e6bbd8-bd14-4785-addd-67c8028cbd22\") " pod="openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw" Nov 22 05:06:14 crc kubenswrapper[4818]: I1122 05:06:14.550146 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5e6bbd8-bd14-4785-addd-67c8028cbd22-bundle\") pod \"f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw\" (UID: \"a5e6bbd8-bd14-4785-addd-67c8028cbd22\") " pod="openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw" Nov 22 05:06:14 crc kubenswrapper[4818]: I1122 05:06:14.571646 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttrst\" (UniqueName: \"kubernetes.io/projected/a5e6bbd8-bd14-4785-addd-67c8028cbd22-kube-api-access-ttrst\") pod \"f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw\" (UID: \"a5e6bbd8-bd14-4785-addd-67c8028cbd22\") " pod="openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw" Nov 22 05:06:14 crc kubenswrapper[4818]: I1122 05:06:14.736743 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw" Nov 22 05:06:15 crc kubenswrapper[4818]: I1122 05:06:15.202975 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw"] Nov 22 05:06:15 crc kubenswrapper[4818]: I1122 05:06:15.299615 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw" event={"ID":"a5e6bbd8-bd14-4785-addd-67c8028cbd22","Type":"ContainerStarted","Data":"02e83594d038499c8ac7aae0f6a751016caa6cc72772bf5f2b295ef083ff1b08"} Nov 22 05:06:16 crc kubenswrapper[4818]: I1122 05:06:16.310060 4818 generic.go:334] "Generic (PLEG): container finished" podID="a5e6bbd8-bd14-4785-addd-67c8028cbd22" containerID="86ee6599dc3e4626c298ccd5e0792d0f9f164ca1825c54ccc01515abf0d56646" exitCode=0 Nov 22 05:06:16 crc kubenswrapper[4818]: I1122 05:06:16.310344 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw" event={"ID":"a5e6bbd8-bd14-4785-addd-67c8028cbd22","Type":"ContainerDied","Data":"86ee6599dc3e4626c298ccd5e0792d0f9f164ca1825c54ccc01515abf0d56646"} Nov 22 05:06:17 crc kubenswrapper[4818]: I1122 05:06:17.319827 4818 generic.go:334] "Generic (PLEG): container finished" podID="a5e6bbd8-bd14-4785-addd-67c8028cbd22" containerID="7599a2f76a592d3d2d970b1727351119ecec68e9b29487fd603960f5da702d6f" exitCode=0 Nov 22 05:06:17 crc kubenswrapper[4818]: I1122 05:06:17.319886 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw" event={"ID":"a5e6bbd8-bd14-4785-addd-67c8028cbd22","Type":"ContainerDied","Data":"7599a2f76a592d3d2d970b1727351119ecec68e9b29487fd603960f5da702d6f"} Nov 22 05:06:18 crc kubenswrapper[4818]: I1122 05:06:18.328162 4818 generic.go:334] "Generic (PLEG): container finished" podID="a5e6bbd8-bd14-4785-addd-67c8028cbd22" containerID="1c7a30bd1831643ba2789632e7d6dcfa1de0905065aa0c320a9deee25274c9a5" exitCode=0 Nov 22 05:06:18 crc kubenswrapper[4818]: I1122 05:06:18.328566 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw" event={"ID":"a5e6bbd8-bd14-4785-addd-67c8028cbd22","Type":"ContainerDied","Data":"1c7a30bd1831643ba2789632e7d6dcfa1de0905065aa0c320a9deee25274c9a5"} Nov 22 05:06:19 crc kubenswrapper[4818]: I1122 05:06:19.596734 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw" Nov 22 05:06:19 crc kubenswrapper[4818]: I1122 05:06:19.616616 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5e6bbd8-bd14-4785-addd-67c8028cbd22-bundle\") pod \"a5e6bbd8-bd14-4785-addd-67c8028cbd22\" (UID: \"a5e6bbd8-bd14-4785-addd-67c8028cbd22\") " Nov 22 05:06:19 crc kubenswrapper[4818]: I1122 05:06:19.616949 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5e6bbd8-bd14-4785-addd-67c8028cbd22-util\") pod \"a5e6bbd8-bd14-4785-addd-67c8028cbd22\" (UID: \"a5e6bbd8-bd14-4785-addd-67c8028cbd22\") " Nov 22 05:06:19 crc kubenswrapper[4818]: I1122 05:06:19.617038 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttrst\" (UniqueName: \"kubernetes.io/projected/a5e6bbd8-bd14-4785-addd-67c8028cbd22-kube-api-access-ttrst\") pod \"a5e6bbd8-bd14-4785-addd-67c8028cbd22\" (UID: \"a5e6bbd8-bd14-4785-addd-67c8028cbd22\") " Nov 22 05:06:19 crc kubenswrapper[4818]: I1122 05:06:19.617470 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5e6bbd8-bd14-4785-addd-67c8028cbd22-bundle" (OuterVolumeSpecName: "bundle") pod "a5e6bbd8-bd14-4785-addd-67c8028cbd22" (UID: "a5e6bbd8-bd14-4785-addd-67c8028cbd22"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:06:19 crc kubenswrapper[4818]: I1122 05:06:19.626587 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5e6bbd8-bd14-4785-addd-67c8028cbd22-kube-api-access-ttrst" (OuterVolumeSpecName: "kube-api-access-ttrst") pod "a5e6bbd8-bd14-4785-addd-67c8028cbd22" (UID: "a5e6bbd8-bd14-4785-addd-67c8028cbd22"). InnerVolumeSpecName "kube-api-access-ttrst". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:06:19 crc kubenswrapper[4818]: I1122 05:06:19.645994 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5e6bbd8-bd14-4785-addd-67c8028cbd22-util" (OuterVolumeSpecName: "util") pod "a5e6bbd8-bd14-4785-addd-67c8028cbd22" (UID: "a5e6bbd8-bd14-4785-addd-67c8028cbd22"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:06:19 crc kubenswrapper[4818]: I1122 05:06:19.721488 4818 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5e6bbd8-bd14-4785-addd-67c8028cbd22-util\") on node \"crc\" DevicePath \"\"" Nov 22 05:06:19 crc kubenswrapper[4818]: I1122 05:06:19.721539 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttrst\" (UniqueName: \"kubernetes.io/projected/a5e6bbd8-bd14-4785-addd-67c8028cbd22-kube-api-access-ttrst\") on node \"crc\" DevicePath \"\"" Nov 22 05:06:19 crc kubenswrapper[4818]: I1122 05:06:19.721551 4818 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5e6bbd8-bd14-4785-addd-67c8028cbd22-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:06:20 crc kubenswrapper[4818]: I1122 05:06:20.345493 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw" event={"ID":"a5e6bbd8-bd14-4785-addd-67c8028cbd22","Type":"ContainerDied","Data":"02e83594d038499c8ac7aae0f6a751016caa6cc72772bf5f2b295ef083ff1b08"} Nov 22 05:06:20 crc kubenswrapper[4818]: I1122 05:06:20.345539 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02e83594d038499c8ac7aae0f6a751016caa6cc72772bf5f2b295ef083ff1b08" Nov 22 05:06:20 crc kubenswrapper[4818]: I1122 05:06:20.345575 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw" Nov 22 05:06:26 crc kubenswrapper[4818]: I1122 05:06:26.993708 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-59b954bb6d-6chhn"] Nov 22 05:06:26 crc kubenswrapper[4818]: E1122 05:06:26.995387 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e6bbd8-bd14-4785-addd-67c8028cbd22" containerName="pull" Nov 22 05:06:26 crc kubenswrapper[4818]: I1122 05:06:26.995466 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e6bbd8-bd14-4785-addd-67c8028cbd22" containerName="pull" Nov 22 05:06:26 crc kubenswrapper[4818]: E1122 05:06:26.995534 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e6bbd8-bd14-4785-addd-67c8028cbd22" containerName="util" Nov 22 05:06:26 crc kubenswrapper[4818]: I1122 05:06:26.995589 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e6bbd8-bd14-4785-addd-67c8028cbd22" containerName="util" Nov 22 05:06:26 crc kubenswrapper[4818]: E1122 05:06:26.995651 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e6bbd8-bd14-4785-addd-67c8028cbd22" containerName="extract" Nov 22 05:06:26 crc kubenswrapper[4818]: I1122 05:06:26.995711 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e6bbd8-bd14-4785-addd-67c8028cbd22" containerName="extract" Nov 22 05:06:26 crc kubenswrapper[4818]: I1122 05:06:26.995864 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5e6bbd8-bd14-4785-addd-67c8028cbd22" containerName="extract" Nov 22 05:06:26 crc kubenswrapper[4818]: I1122 05:06:26.996635 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-59b954bb6d-6chhn" Nov 22 05:06:27 crc kubenswrapper[4818]: I1122 05:06:26.999515 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-hw9bn" Nov 22 05:06:27 crc kubenswrapper[4818]: I1122 05:06:27.020279 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-59b954bb6d-6chhn"] Nov 22 05:06:27 crc kubenswrapper[4818]: I1122 05:06:27.020686 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkw7s\" (UniqueName: \"kubernetes.io/projected/04e2d18d-98fc-4fa8-9a1f-1742f497d737-kube-api-access-fkw7s\") pod \"openstack-operator-controller-operator-59b954bb6d-6chhn\" (UID: \"04e2d18d-98fc-4fa8-9a1f-1742f497d737\") " pod="openstack-operators/openstack-operator-controller-operator-59b954bb6d-6chhn" Nov 22 05:06:27 crc kubenswrapper[4818]: I1122 05:06:27.121548 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkw7s\" (UniqueName: \"kubernetes.io/projected/04e2d18d-98fc-4fa8-9a1f-1742f497d737-kube-api-access-fkw7s\") pod \"openstack-operator-controller-operator-59b954bb6d-6chhn\" (UID: \"04e2d18d-98fc-4fa8-9a1f-1742f497d737\") " pod="openstack-operators/openstack-operator-controller-operator-59b954bb6d-6chhn" Nov 22 05:06:27 crc kubenswrapper[4818]: I1122 05:06:27.141979 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkw7s\" (UniqueName: \"kubernetes.io/projected/04e2d18d-98fc-4fa8-9a1f-1742f497d737-kube-api-access-fkw7s\") pod \"openstack-operator-controller-operator-59b954bb6d-6chhn\" (UID: \"04e2d18d-98fc-4fa8-9a1f-1742f497d737\") " pod="openstack-operators/openstack-operator-controller-operator-59b954bb6d-6chhn" Nov 22 05:06:27 crc kubenswrapper[4818]: I1122 05:06:27.314750 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-59b954bb6d-6chhn" Nov 22 05:06:27 crc kubenswrapper[4818]: I1122 05:06:27.778026 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-59b954bb6d-6chhn"] Nov 22 05:06:28 crc kubenswrapper[4818]: I1122 05:06:28.395274 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-59b954bb6d-6chhn" event={"ID":"04e2d18d-98fc-4fa8-9a1f-1742f497d737","Type":"ContainerStarted","Data":"aa6f71a7e5590da82dd990a0514182e89b8109d2e52967dcc689a6815bbf925b"} Nov 22 05:06:31 crc kubenswrapper[4818]: I1122 05:06:31.415220 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-59b954bb6d-6chhn" event={"ID":"04e2d18d-98fc-4fa8-9a1f-1742f497d737","Type":"ContainerStarted","Data":"a71b8116650188bc68512ec3639affe8bbcf65a8af2829d87627fa56151e1e3c"} Nov 22 05:06:38 crc kubenswrapper[4818]: I1122 05:06:38.465913 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-59b954bb6d-6chhn" event={"ID":"04e2d18d-98fc-4fa8-9a1f-1742f497d737","Type":"ContainerStarted","Data":"c0ee7a125a6a8e9266f20595e5432c293ffb1d652f984292f895bb2071c28657"} Nov 22 05:06:38 crc kubenswrapper[4818]: I1122 05:06:38.467324 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-59b954bb6d-6chhn" Nov 22 05:06:38 crc kubenswrapper[4818]: I1122 05:06:38.468787 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-59b954bb6d-6chhn" Nov 22 05:06:38 crc kubenswrapper[4818]: I1122 05:06:38.503658 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-59b954bb6d-6chhn" podStartSLOduration=2.598425823 podStartE2EDuration="12.503641195s" podCreationTimestamp="2025-11-22 05:06:26 +0000 UTC" firstStartedPulling="2025-11-22 05:06:27.796556174 +0000 UTC m=+1140.370972701" lastFinishedPulling="2025-11-22 05:06:37.701771536 +0000 UTC m=+1150.276188073" observedRunningTime="2025-11-22 05:06:38.502616938 +0000 UTC m=+1151.077033475" watchObservedRunningTime="2025-11-22 05:06:38.503641195 +0000 UTC m=+1151.078057732" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.734807 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-8vknn"] Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.737688 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-8vknn" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.740106 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-8p2cm" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.748904 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thdjq\" (UniqueName: \"kubernetes.io/projected/69877edd-8b97-4838-bc41-c087fc4deb7e-kube-api-access-thdjq\") pod \"barbican-operator-controller-manager-75fb479bcc-8vknn\" (UID: \"69877edd-8b97-4838-bc41-c087fc4deb7e\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-8vknn" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.749358 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-b2lp5"] Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.751885 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-b2lp5" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.756720 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-wvlkt" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.764362 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-8vknn"] Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.784636 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-b2lp5"] Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.807570 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-jttdm"] Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.809853 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-jttdm" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.817435 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-bkwnl" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.860396 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thdjq\" (UniqueName: \"kubernetes.io/projected/69877edd-8b97-4838-bc41-c087fc4deb7e-kube-api-access-thdjq\") pod \"barbican-operator-controller-manager-75fb479bcc-8vknn\" (UID: \"69877edd-8b97-4838-bc41-c087fc4deb7e\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-8vknn" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.864270 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-cl4tl"] Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.865348 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-cl4tl" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.867495 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-mrq7q" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.896336 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-jttdm"] Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.926310 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-cl4tl"] Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.934691 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-qt9nt"] Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.953377 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thdjq\" (UniqueName: \"kubernetes.io/projected/69877edd-8b97-4838-bc41-c087fc4deb7e-kube-api-access-thdjq\") pod \"barbican-operator-controller-manager-75fb479bcc-8vknn\" (UID: \"69877edd-8b97-4838-bc41-c087fc4deb7e\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-8vknn" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.954751 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-7875d8bb94-dn8xx"] Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.955660 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-4zl42"] Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.956437 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-qt9nt" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.958356 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-dn8xx" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.959948 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-4zl42"] Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.960194 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4zl42" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.963636 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-qt9nt"] Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.967430 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.967516 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-mkgt8" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.968022 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-bgnr7" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.971754 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-mgts5" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.978972 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7875d8bb94-dn8xx"] Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.989153 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4ggr\" (UniqueName: \"kubernetes.io/projected/05bab737-6cda-4891-a014-202ab05b141c-kube-api-access-w4ggr\") pod \"horizon-operator-controller-manager-598f69df5d-4zl42\" (UID: \"05bab737-6cda-4891-a014-202ab05b141c\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4zl42" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.989203 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8cqf\" (UniqueName: \"kubernetes.io/projected/27ec8b4f-e5ec-483a-ada8-6a766b843b73-kube-api-access-r8cqf\") pod \"cinder-operator-controller-manager-6498cbf48f-b2lp5\" (UID: \"27ec8b4f-e5ec-483a-ada8-6a766b843b73\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-b2lp5" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.989334 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhg87\" (UniqueName: \"kubernetes.io/projected/7389ff62-006d-4d11-8237-e57c9279b0f3-kube-api-access-dhg87\") pod \"infra-operator-controller-manager-7875d8bb94-dn8xx\" (UID: \"7389ff62-006d-4d11-8237-e57c9279b0f3\") " pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-dn8xx" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.989430 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjrg4\" (UniqueName: \"kubernetes.io/projected/b9a85b40-dd76-45b2-bd41-ecb5be7387da-kube-api-access-vjrg4\") pod \"designate-operator-controller-manager-767ccfd65f-jttdm\" (UID: \"b9a85b40-dd76-45b2-bd41-ecb5be7387da\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-jttdm" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.989473 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t67pg\" (UniqueName: \"kubernetes.io/projected/10da5c26-c5b3-4477-a4d4-a5c98741dfb1-kube-api-access-t67pg\") pod \"heat-operator-controller-manager-56f54d6746-qt9nt\" (UID: \"10da5c26-c5b3-4477-a4d4-a5c98741dfb1\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-qt9nt" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.989499 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k4fw\" (UniqueName: \"kubernetes.io/projected/4d915797-cd52-4f07-9028-4f80411fe65b-kube-api-access-4k4fw\") pod \"glance-operator-controller-manager-7969689c84-cl4tl\" (UID: \"4d915797-cd52-4f07-9028-4f80411fe65b\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-cl4tl" Nov 22 05:07:13 crc kubenswrapper[4818]: I1122 05:07:13.989548 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7389ff62-006d-4d11-8237-e57c9279b0f3-cert\") pod \"infra-operator-controller-manager-7875d8bb94-dn8xx\" (UID: \"7389ff62-006d-4d11-8237-e57c9279b0f3\") " pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-dn8xx" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.002712 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-ncfxv"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.003699 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ncfxv" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.005840 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-scjq7" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.014744 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-pqd4c"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.016059 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-pqd4c" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.022113 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-vc2qn" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.035306 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-ncfxv"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.065522 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-85bf8d6cb7-lzmmp"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.066504 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-85bf8d6cb7-lzmmp" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.073364 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-8vknn" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.073623 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-fbjqf" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.081345 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-85bf8d6cb7-lzmmp"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.090733 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhg87\" (UniqueName: \"kubernetes.io/projected/7389ff62-006d-4d11-8237-e57c9279b0f3-kube-api-access-dhg87\") pod \"infra-operator-controller-manager-7875d8bb94-dn8xx\" (UID: \"7389ff62-006d-4d11-8237-e57c9279b0f3\") " pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-dn8xx" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.091072 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjrg4\" (UniqueName: \"kubernetes.io/projected/b9a85b40-dd76-45b2-bd41-ecb5be7387da-kube-api-access-vjrg4\") pod \"designate-operator-controller-manager-767ccfd65f-jttdm\" (UID: \"b9a85b40-dd76-45b2-bd41-ecb5be7387da\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-jttdm" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.091099 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t67pg\" (UniqueName: \"kubernetes.io/projected/10da5c26-c5b3-4477-a4d4-a5c98741dfb1-kube-api-access-t67pg\") pod \"heat-operator-controller-manager-56f54d6746-qt9nt\" (UID: \"10da5c26-c5b3-4477-a4d4-a5c98741dfb1\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-qt9nt" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.091117 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4k4fw\" (UniqueName: \"kubernetes.io/projected/4d915797-cd52-4f07-9028-4f80411fe65b-kube-api-access-4k4fw\") pod \"glance-operator-controller-manager-7969689c84-cl4tl\" (UID: \"4d915797-cd52-4f07-9028-4f80411fe65b\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-cl4tl" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.091145 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7389ff62-006d-4d11-8237-e57c9279b0f3-cert\") pod \"infra-operator-controller-manager-7875d8bb94-dn8xx\" (UID: \"7389ff62-006d-4d11-8237-e57c9279b0f3\") " pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-dn8xx" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.091167 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8cqf\" (UniqueName: \"kubernetes.io/projected/27ec8b4f-e5ec-483a-ada8-6a766b843b73-kube-api-access-r8cqf\") pod \"cinder-operator-controller-manager-6498cbf48f-b2lp5\" (UID: \"27ec8b4f-e5ec-483a-ada8-6a766b843b73\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-b2lp5" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.091182 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4ggr\" (UniqueName: \"kubernetes.io/projected/05bab737-6cda-4891-a014-202ab05b141c-kube-api-access-w4ggr\") pod \"horizon-operator-controller-manager-598f69df5d-4zl42\" (UID: \"05bab737-6cda-4891-a014-202ab05b141c\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4zl42" Nov 22 05:07:14 crc kubenswrapper[4818]: E1122 05:07:14.091670 4818 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 22 05:07:14 crc kubenswrapper[4818]: E1122 05:07:14.091719 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7389ff62-006d-4d11-8237-e57c9279b0f3-cert podName:7389ff62-006d-4d11-8237-e57c9279b0f3 nodeName:}" failed. No retries permitted until 2025-11-22 05:07:14.591700691 +0000 UTC m=+1187.166117218 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7389ff62-006d-4d11-8237-e57c9279b0f3-cert") pod "infra-operator-controller-manager-7875d8bb94-dn8xx" (UID: "7389ff62-006d-4d11-8237-e57c9279b0f3") : secret "infra-operator-webhook-server-cert" not found Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.101339 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-pqd4c"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.127902 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t67pg\" (UniqueName: \"kubernetes.io/projected/10da5c26-c5b3-4477-a4d4-a5c98741dfb1-kube-api-access-t67pg\") pod \"heat-operator-controller-manager-56f54d6746-qt9nt\" (UID: \"10da5c26-c5b3-4477-a4d4-a5c98741dfb1\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-qt9nt" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.133236 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4k4fw\" (UniqueName: \"kubernetes.io/projected/4d915797-cd52-4f07-9028-4f80411fe65b-kube-api-access-4k4fw\") pod \"glance-operator-controller-manager-7969689c84-cl4tl\" (UID: \"4d915797-cd52-4f07-9028-4f80411fe65b\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-cl4tl" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.134754 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhg87\" (UniqueName: \"kubernetes.io/projected/7389ff62-006d-4d11-8237-e57c9279b0f3-kube-api-access-dhg87\") pod \"infra-operator-controller-manager-7875d8bb94-dn8xx\" (UID: \"7389ff62-006d-4d11-8237-e57c9279b0f3\") " pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-dn8xx" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.134806 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-8wk6c"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.146936 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-8wk6c" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.157192 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4ggr\" (UniqueName: \"kubernetes.io/projected/05bab737-6cda-4891-a014-202ab05b141c-kube-api-access-w4ggr\") pod \"horizon-operator-controller-manager-598f69df5d-4zl42\" (UID: \"05bab737-6cda-4891-a014-202ab05b141c\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4zl42" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.157589 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-w5z52" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.158115 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-rf9vq"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.159044 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-rf9vq" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.161684 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-2ghxn" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.165967 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-lx77w"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.166070 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjrg4\" (UniqueName: \"kubernetes.io/projected/b9a85b40-dd76-45b2-bd41-ecb5be7387da-kube-api-access-vjrg4\") pod \"designate-operator-controller-manager-767ccfd65f-jttdm\" (UID: \"b9a85b40-dd76-45b2-bd41-ecb5be7387da\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-jttdm" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.167275 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-lx77w" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.167962 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8cqf\" (UniqueName: \"kubernetes.io/projected/27ec8b4f-e5ec-483a-ada8-6a766b843b73-kube-api-access-r8cqf\") pod \"cinder-operator-controller-manager-6498cbf48f-b2lp5\" (UID: \"27ec8b4f-e5ec-483a-ada8-6a766b843b73\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-b2lp5" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.170244 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-ccd8t" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.173710 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-jttdm" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.174307 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-8wk6c"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.180972 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-xtplb"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.182195 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xtplb" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.184227 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-df97x" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.186170 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-lx77w"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.187751 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-cl4tl" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.191051 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-rf9vq"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.194521 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx8fs\" (UniqueName: \"kubernetes.io/projected/d6b59804-efc5-4f60-b9d3-bf0e998c8f89-kube-api-access-lx8fs\") pod \"keystone-operator-controller-manager-7454b96578-pqd4c\" (UID: \"d6b59804-efc5-4f60-b9d3-bf0e998c8f89\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-pqd4c" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.194598 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxwc5\" (UniqueName: \"kubernetes.io/projected/4838407d-e782-4aab-8fbf-861f3ba9019a-kube-api-access-wxwc5\") pod \"manila-operator-controller-manager-85bf8d6cb7-lzmmp\" (UID: \"4838407d-e782-4aab-8fbf-861f3ba9019a\") " pod="openstack-operators/manila-operator-controller-manager-85bf8d6cb7-lzmmp" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.194654 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g445z\" (UniqueName: \"kubernetes.io/projected/c31258b3-6cac-41e8-a350-1ea7b360b848-kube-api-access-g445z\") pod \"ironic-operator-controller-manager-99b499f4-ncfxv\" (UID: \"c31258b3-6cac-41e8-a350-1ea7b360b848\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ncfxv" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.200775 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-xtplb"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.206320 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-n54fq"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.207272 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-n54fq" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.217393 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-8dhvf"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.218422 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-8dhvf" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.222721 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.222875 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-46jgm" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.223051 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-7cz92" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.236380 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-8dhvf"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.283326 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-56phb"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.284845 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-56phb" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.295236 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-qt9nt" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.296905 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxwc5\" (UniqueName: \"kubernetes.io/projected/4838407d-e782-4aab-8fbf-861f3ba9019a-kube-api-access-wxwc5\") pod \"manila-operator-controller-manager-85bf8d6cb7-lzmmp\" (UID: \"4838407d-e782-4aab-8fbf-861f3ba9019a\") " pod="openstack-operators/manila-operator-controller-manager-85bf8d6cb7-lzmmp" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.296957 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xs7j\" (UniqueName: \"kubernetes.io/projected/68668b7f-5fc3-4774-8f0d-3eae71839d52-kube-api-access-5xs7j\") pod \"neutron-operator-controller-manager-78bd47f458-xtplb\" (UID: \"68668b7f-5fc3-4774-8f0d-3eae71839d52\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xtplb" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.296986 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw67r\" (UniqueName: \"kubernetes.io/projected/96a59329-8f96-43a5-9abc-fa8d1d7009fb-kube-api-access-dw67r\") pod \"mariadb-operator-controller-manager-54b5986bb8-8wk6c\" (UID: \"96a59329-8f96-43a5-9abc-fa8d1d7009fb\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-8wk6c" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.297009 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pstmm\" (UniqueName: \"kubernetes.io/projected/06ec2990-cd45-4808-9748-2a1d75c5015f-kube-api-access-pstmm\") pod \"ovn-operator-controller-manager-54fc5f65b7-8dhvf\" (UID: \"06ec2990-cd45-4808-9748-2a1d75c5015f\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-8dhvf" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.297066 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s985v\" (UniqueName: \"kubernetes.io/projected/9797ea62-91f5-4b4f-a9f0-27911cf629e4-kube-api-access-s985v\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-n54fq\" (UID: \"9797ea62-91f5-4b4f-a9f0-27911cf629e4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-n54fq" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.297099 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g445z\" (UniqueName: \"kubernetes.io/projected/c31258b3-6cac-41e8-a350-1ea7b360b848-kube-api-access-g445z\") pod \"ironic-operator-controller-manager-99b499f4-ncfxv\" (UID: \"c31258b3-6cac-41e8-a350-1ea7b360b848\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ncfxv" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.297115 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9797ea62-91f5-4b4f-a9f0-27911cf629e4-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-n54fq\" (UID: \"9797ea62-91f5-4b4f-a9f0-27911cf629e4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-n54fq" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.297142 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx8fs\" (UniqueName: \"kubernetes.io/projected/d6b59804-efc5-4f60-b9d3-bf0e998c8f89-kube-api-access-lx8fs\") pod \"keystone-operator-controller-manager-7454b96578-pqd4c\" (UID: \"d6b59804-efc5-4f60-b9d3-bf0e998c8f89\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-pqd4c" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.297179 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxr94\" (UniqueName: \"kubernetes.io/projected/10d7f6df-ae52-4dab-b49d-764d95f77935-kube-api-access-wxr94\") pod \"octavia-operator-controller-manager-54cfbf4c7d-rf9vq\" (UID: \"10d7f6df-ae52-4dab-b49d-764d95f77935\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-rf9vq" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.297204 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmxgx\" (UniqueName: \"kubernetes.io/projected/531548a2-b8d3-4a2a-a6e9-1bb2af1891d8-kube-api-access-dmxgx\") pod \"nova-operator-controller-manager-cfbb9c588-lx77w\" (UID: \"531548a2-b8d3-4a2a-a6e9-1bb2af1891d8\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-lx77w" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.297224 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqlnl\" (UniqueName: \"kubernetes.io/projected/786ab73e-012b-47f6-b033-9fa9aab6d6cb-kube-api-access-jqlnl\") pod \"placement-operator-controller-manager-5b797b8dff-56phb\" (UID: \"786ab73e-012b-47f6-b033-9fa9aab6d6cb\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-56phb" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.297893 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-gktn8" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.356380 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4zl42" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.377075 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx8fs\" (UniqueName: \"kubernetes.io/projected/d6b59804-efc5-4f60-b9d3-bf0e998c8f89-kube-api-access-lx8fs\") pod \"keystone-operator-controller-manager-7454b96578-pqd4c\" (UID: \"d6b59804-efc5-4f60-b9d3-bf0e998c8f89\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-pqd4c" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.393733 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-pqd4c" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.394642 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-b2lp5" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.399073 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xs7j\" (UniqueName: \"kubernetes.io/projected/68668b7f-5fc3-4774-8f0d-3eae71839d52-kube-api-access-5xs7j\") pod \"neutron-operator-controller-manager-78bd47f458-xtplb\" (UID: \"68668b7f-5fc3-4774-8f0d-3eae71839d52\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xtplb" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.399136 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw67r\" (UniqueName: \"kubernetes.io/projected/96a59329-8f96-43a5-9abc-fa8d1d7009fb-kube-api-access-dw67r\") pod \"mariadb-operator-controller-manager-54b5986bb8-8wk6c\" (UID: \"96a59329-8f96-43a5-9abc-fa8d1d7009fb\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-8wk6c" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.399164 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pstmm\" (UniqueName: \"kubernetes.io/projected/06ec2990-cd45-4808-9748-2a1d75c5015f-kube-api-access-pstmm\") pod \"ovn-operator-controller-manager-54fc5f65b7-8dhvf\" (UID: \"06ec2990-cd45-4808-9748-2a1d75c5015f\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-8dhvf" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.399199 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s985v\" (UniqueName: \"kubernetes.io/projected/9797ea62-91f5-4b4f-a9f0-27911cf629e4-kube-api-access-s985v\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-n54fq\" (UID: \"9797ea62-91f5-4b4f-a9f0-27911cf629e4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-n54fq" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.399241 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9797ea62-91f5-4b4f-a9f0-27911cf629e4-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-n54fq\" (UID: \"9797ea62-91f5-4b4f-a9f0-27911cf629e4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-n54fq" Nov 22 05:07:14 crc kubenswrapper[4818]: E1122 05:07:14.421488 4818 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 22 05:07:14 crc kubenswrapper[4818]: E1122 05:07:14.423356 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9797ea62-91f5-4b4f-a9f0-27911cf629e4-cert podName:9797ea62-91f5-4b4f-a9f0-27911cf629e4 nodeName:}" failed. No retries permitted until 2025-11-22 05:07:14.923335522 +0000 UTC m=+1187.497752139 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9797ea62-91f5-4b4f-a9f0-27911cf629e4-cert") pod "openstack-baremetal-operator-controller-manager-8c7444f48-n54fq" (UID: "9797ea62-91f5-4b4f-a9f0-27911cf629e4") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.464014 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxr94\" (UniqueName: \"kubernetes.io/projected/10d7f6df-ae52-4dab-b49d-764d95f77935-kube-api-access-wxr94\") pod \"octavia-operator-controller-manager-54cfbf4c7d-rf9vq\" (UID: \"10d7f6df-ae52-4dab-b49d-764d95f77935\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-rf9vq" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.464113 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmxgx\" (UniqueName: \"kubernetes.io/projected/531548a2-b8d3-4a2a-a6e9-1bb2af1891d8-kube-api-access-dmxgx\") pod \"nova-operator-controller-manager-cfbb9c588-lx77w\" (UID: \"531548a2-b8d3-4a2a-a6e9-1bb2af1891d8\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-lx77w" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.464144 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqlnl\" (UniqueName: \"kubernetes.io/projected/786ab73e-012b-47f6-b033-9fa9aab6d6cb-kube-api-access-jqlnl\") pod \"placement-operator-controller-manager-5b797b8dff-56phb\" (UID: \"786ab73e-012b-47f6-b033-9fa9aab6d6cb\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-56phb" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.426345 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-n54fq"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.464724 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-56phb"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.464743 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-vkj22"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.468284 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxwc5\" (UniqueName: \"kubernetes.io/projected/4838407d-e782-4aab-8fbf-861f3ba9019a-kube-api-access-wxwc5\") pod \"manila-operator-controller-manager-85bf8d6cb7-lzmmp\" (UID: \"4838407d-e782-4aab-8fbf-861f3ba9019a\") " pod="openstack-operators/manila-operator-controller-manager-85bf8d6cb7-lzmmp" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.470602 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-vkj22" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.458382 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g445z\" (UniqueName: \"kubernetes.io/projected/c31258b3-6cac-41e8-a350-1ea7b360b848-kube-api-access-g445z\") pod \"ironic-operator-controller-manager-99b499f4-ncfxv\" (UID: \"c31258b3-6cac-41e8-a350-1ea7b360b848\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ncfxv" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.474090 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-hbp8l" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.543293 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pstmm\" (UniqueName: \"kubernetes.io/projected/06ec2990-cd45-4808-9748-2a1d75c5015f-kube-api-access-pstmm\") pod \"ovn-operator-controller-manager-54fc5f65b7-8dhvf\" (UID: \"06ec2990-cd45-4808-9748-2a1d75c5015f\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-8dhvf" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.546246 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw67r\" (UniqueName: \"kubernetes.io/projected/96a59329-8f96-43a5-9abc-fa8d1d7009fb-kube-api-access-dw67r\") pod \"mariadb-operator-controller-manager-54b5986bb8-8wk6c\" (UID: \"96a59329-8f96-43a5-9abc-fa8d1d7009fb\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-8wk6c" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.546971 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s985v\" (UniqueName: \"kubernetes.io/projected/9797ea62-91f5-4b4f-a9f0-27911cf629e4-kube-api-access-s985v\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-n54fq\" (UID: \"9797ea62-91f5-4b4f-a9f0-27911cf629e4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-n54fq" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.550700 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-85bf8d6cb7-lzmmp" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.563134 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-8wk6c" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.565127 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xs7j\" (UniqueName: \"kubernetes.io/projected/68668b7f-5fc3-4774-8f0d-3eae71839d52-kube-api-access-5xs7j\") pod \"neutron-operator-controller-manager-78bd47f458-xtplb\" (UID: \"68668b7f-5fc3-4774-8f0d-3eae71839d52\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xtplb" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.566601 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-996gn\" (UniqueName: \"kubernetes.io/projected/f04a965b-f5f4-45af-843f-862498956853-kube-api-access-996gn\") pod \"swift-operator-controller-manager-d656998f4-vkj22\" (UID: \"f04a965b-f5f4-45af-843f-862498956853\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-vkj22" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.568400 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmxgx\" (UniqueName: \"kubernetes.io/projected/531548a2-b8d3-4a2a-a6e9-1bb2af1891d8-kube-api-access-dmxgx\") pod \"nova-operator-controller-manager-cfbb9c588-lx77w\" (UID: \"531548a2-b8d3-4a2a-a6e9-1bb2af1891d8\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-lx77w" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.569226 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqlnl\" (UniqueName: \"kubernetes.io/projected/786ab73e-012b-47f6-b033-9fa9aab6d6cb-kube-api-access-jqlnl\") pod \"placement-operator-controller-manager-5b797b8dff-56phb\" (UID: \"786ab73e-012b-47f6-b033-9fa9aab6d6cb\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-56phb" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.576906 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxr94\" (UniqueName: \"kubernetes.io/projected/10d7f6df-ae52-4dab-b49d-764d95f77935-kube-api-access-wxr94\") pod \"octavia-operator-controller-manager-54cfbf4c7d-rf9vq\" (UID: \"10d7f6df-ae52-4dab-b49d-764d95f77935\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-rf9vq" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.585105 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xtplb" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.591662 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-vkj22"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.591748 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-cjphz"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.592967 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-cjphz" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.599821 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-tblnr" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.609120 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-cjphz"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.642191 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-pb45n"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.644409 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-pb45n" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.648429 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-8dhvf" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.649221 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-n5k4h" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.651492 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-pb45n"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.667534 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-996gn\" (UniqueName: \"kubernetes.io/projected/f04a965b-f5f4-45af-843f-862498956853-kube-api-access-996gn\") pod \"swift-operator-controller-manager-d656998f4-vkj22\" (UID: \"f04a965b-f5f4-45af-843f-862498956853\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-vkj22" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.667619 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqhlq\" (UniqueName: \"kubernetes.io/projected/e342ff02-d6d9-4be0-b97f-87dea6653a61-kube-api-access-vqhlq\") pod \"test-operator-controller-manager-b4c496f69-pb45n\" (UID: \"e342ff02-d6d9-4be0-b97f-87dea6653a61\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-pb45n" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.667727 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gshng\" (UniqueName: \"kubernetes.io/projected/bba055bd-511b-4af6-bc94-4cadf7a84853-kube-api-access-gshng\") pod \"telemetry-operator-controller-manager-6d4bf84b58-cjphz\" (UID: \"bba055bd-511b-4af6-bc94-4cadf7a84853\") " pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-cjphz" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.667760 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7389ff62-006d-4d11-8237-e57c9279b0f3-cert\") pod \"infra-operator-controller-manager-7875d8bb94-dn8xx\" (UID: \"7389ff62-006d-4d11-8237-e57c9279b0f3\") " pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-dn8xx" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.672405 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7389ff62-006d-4d11-8237-e57c9279b0f3-cert\") pod \"infra-operator-controller-manager-7875d8bb94-dn8xx\" (UID: \"7389ff62-006d-4d11-8237-e57c9279b0f3\") " pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-dn8xx" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.675549 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ncfxv" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.676651 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-sqvdq"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.680318 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-sqvdq" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.685203 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-gdlz7" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.686189 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-sqvdq"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.688197 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-56phb" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.689490 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-996gn\" (UniqueName: \"kubernetes.io/projected/f04a965b-f5f4-45af-843f-862498956853-kube-api-access-996gn\") pod \"swift-operator-controller-manager-d656998f4-vkj22\" (UID: \"f04a965b-f5f4-45af-843f-862498956853\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-vkj22" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.708402 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6b57b64d88-pnqbd"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.709946 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6b57b64d88-pnqbd" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.713282 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-txzd8" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.713454 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.714043 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-rf9vq" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.717992 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-vkj22" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.738321 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6b57b64d88-pnqbd"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.757764 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d6957"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.768911 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d6957" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.771281 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gshng\" (UniqueName: \"kubernetes.io/projected/bba055bd-511b-4af6-bc94-4cadf7a84853-kube-api-access-gshng\") pod \"telemetry-operator-controller-manager-6d4bf84b58-cjphz\" (UID: \"bba055bd-511b-4af6-bc94-4cadf7a84853\") " pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-cjphz" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.771325 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66zvd\" (UniqueName: \"kubernetes.io/projected/2b80bb62-ec21-44c6-91e2-b422b9813255-kube-api-access-66zvd\") pod \"watcher-operator-controller-manager-8c6448b9f-sqvdq\" (UID: \"2b80bb62-ec21-44c6-91e2-b422b9813255\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-sqvdq" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.771372 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rx2jf\" (UniqueName: \"kubernetes.io/projected/44f3f34b-8fa2-479c-a876-947fab621708-kube-api-access-rx2jf\") pod \"openstack-operator-controller-manager-6b57b64d88-pnqbd\" (UID: \"44f3f34b-8fa2-479c-a876-947fab621708\") " pod="openstack-operators/openstack-operator-controller-manager-6b57b64d88-pnqbd" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.771435 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/44f3f34b-8fa2-479c-a876-947fab621708-cert\") pod \"openstack-operator-controller-manager-6b57b64d88-pnqbd\" (UID: \"44f3f34b-8fa2-479c-a876-947fab621708\") " pod="openstack-operators/openstack-operator-controller-manager-6b57b64d88-pnqbd" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.771455 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqhlq\" (UniqueName: \"kubernetes.io/projected/e342ff02-d6d9-4be0-b97f-87dea6653a61-kube-api-access-vqhlq\") pod \"test-operator-controller-manager-b4c496f69-pb45n\" (UID: \"e342ff02-d6d9-4be0-b97f-87dea6653a61\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-pb45n" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.786990 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d6957"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.792831 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-mkwh9" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.795240 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-8vknn"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.816765 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gshng\" (UniqueName: \"kubernetes.io/projected/bba055bd-511b-4af6-bc94-4cadf7a84853-kube-api-access-gshng\") pod \"telemetry-operator-controller-manager-6d4bf84b58-cjphz\" (UID: \"bba055bd-511b-4af6-bc94-4cadf7a84853\") " pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-cjphz" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.818955 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqhlq\" (UniqueName: \"kubernetes.io/projected/e342ff02-d6d9-4be0-b97f-87dea6653a61-kube-api-access-vqhlq\") pod \"test-operator-controller-manager-b4c496f69-pb45n\" (UID: \"e342ff02-d6d9-4be0-b97f-87dea6653a61\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-pb45n" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.848477 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-lx77w" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.880820 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr9dc\" (UniqueName: \"kubernetes.io/projected/62c4e1cf-72fb-4a9b-88a1-350c1e20f360-kube-api-access-xr9dc\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-d6957\" (UID: \"62c4e1cf-72fb-4a9b-88a1-350c1e20f360\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d6957" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.880919 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/44f3f34b-8fa2-479c-a876-947fab621708-cert\") pod \"openstack-operator-controller-manager-6b57b64d88-pnqbd\" (UID: \"44f3f34b-8fa2-479c-a876-947fab621708\") " pod="openstack-operators/openstack-operator-controller-manager-6b57b64d88-pnqbd" Nov 22 05:07:14 crc kubenswrapper[4818]: E1122 05:07:14.881034 4818 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 22 05:07:14 crc kubenswrapper[4818]: E1122 05:07:14.881309 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/44f3f34b-8fa2-479c-a876-947fab621708-cert podName:44f3f34b-8fa2-479c-a876-947fab621708 nodeName:}" failed. No retries permitted until 2025-11-22 05:07:15.381096677 +0000 UTC m=+1187.955513214 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/44f3f34b-8fa2-479c-a876-947fab621708-cert") pod "openstack-operator-controller-manager-6b57b64d88-pnqbd" (UID: "44f3f34b-8fa2-479c-a876-947fab621708") : secret "webhook-server-cert" not found Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.881871 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66zvd\" (UniqueName: \"kubernetes.io/projected/2b80bb62-ec21-44c6-91e2-b422b9813255-kube-api-access-66zvd\") pod \"watcher-operator-controller-manager-8c6448b9f-sqvdq\" (UID: \"2b80bb62-ec21-44c6-91e2-b422b9813255\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-sqvdq" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.881924 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rx2jf\" (UniqueName: \"kubernetes.io/projected/44f3f34b-8fa2-479c-a876-947fab621708-kube-api-access-rx2jf\") pod \"openstack-operator-controller-manager-6b57b64d88-pnqbd\" (UID: \"44f3f34b-8fa2-479c-a876-947fab621708\") " pod="openstack-operators/openstack-operator-controller-manager-6b57b64d88-pnqbd" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.903056 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66zvd\" (UniqueName: \"kubernetes.io/projected/2b80bb62-ec21-44c6-91e2-b422b9813255-kube-api-access-66zvd\") pod \"watcher-operator-controller-manager-8c6448b9f-sqvdq\" (UID: \"2b80bb62-ec21-44c6-91e2-b422b9813255\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-sqvdq" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.905317 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rx2jf\" (UniqueName: \"kubernetes.io/projected/44f3f34b-8fa2-479c-a876-947fab621708-kube-api-access-rx2jf\") pod \"openstack-operator-controller-manager-6b57b64d88-pnqbd\" (UID: \"44f3f34b-8fa2-479c-a876-947fab621708\") " pod="openstack-operators/openstack-operator-controller-manager-6b57b64d88-pnqbd" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.933019 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-dn8xx" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.964725 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-jttdm"] Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.985598 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr9dc\" (UniqueName: \"kubernetes.io/projected/62c4e1cf-72fb-4a9b-88a1-350c1e20f360-kube-api-access-xr9dc\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-d6957\" (UID: \"62c4e1cf-72fb-4a9b-88a1-350c1e20f360\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d6957" Nov 22 05:07:14 crc kubenswrapper[4818]: I1122 05:07:14.987892 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9797ea62-91f5-4b4f-a9f0-27911cf629e4-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-n54fq\" (UID: \"9797ea62-91f5-4b4f-a9f0-27911cf629e4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-n54fq" Nov 22 05:07:14 crc kubenswrapper[4818]: E1122 05:07:14.988112 4818 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 22 05:07:14 crc kubenswrapper[4818]: E1122 05:07:14.988171 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9797ea62-91f5-4b4f-a9f0-27911cf629e4-cert podName:9797ea62-91f5-4b4f-a9f0-27911cf629e4 nodeName:}" failed. No retries permitted until 2025-11-22 05:07:15.988152762 +0000 UTC m=+1188.562569279 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9797ea62-91f5-4b4f-a9f0-27911cf629e4-cert") pod "openstack-baremetal-operator-controller-manager-8c7444f48-n54fq" (UID: "9797ea62-91f5-4b4f-a9f0-27911cf629e4") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.006610 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-cl4tl"] Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.025723 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr9dc\" (UniqueName: \"kubernetes.io/projected/62c4e1cf-72fb-4a9b-88a1-350c1e20f360-kube-api-access-xr9dc\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-d6957\" (UID: \"62c4e1cf-72fb-4a9b-88a1-350c1e20f360\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d6957" Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.051011 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-cjphz" Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.075453 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-pb45n" Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.101641 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-sqvdq" Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.325056 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d6957" Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.402872 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/44f3f34b-8fa2-479c-a876-947fab621708-cert\") pod \"openstack-operator-controller-manager-6b57b64d88-pnqbd\" (UID: \"44f3f34b-8fa2-479c-a876-947fab621708\") " pod="openstack-operators/openstack-operator-controller-manager-6b57b64d88-pnqbd" Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.406813 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/44f3f34b-8fa2-479c-a876-947fab621708-cert\") pod \"openstack-operator-controller-manager-6b57b64d88-pnqbd\" (UID: \"44f3f34b-8fa2-479c-a876-947fab621708\") " pod="openstack-operators/openstack-operator-controller-manager-6b57b64d88-pnqbd" Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.466832 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6b57b64d88-pnqbd" Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.555518 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-xtplb"] Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.562606 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-4zl42"] Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.580711 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-qt9nt"] Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.696929 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-85bf8d6cb7-lzmmp"] Nov 22 05:07:15 crc kubenswrapper[4818]: W1122 05:07:15.697370 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4838407d_e782_4aab_8fbf_861f3ba9019a.slice/crio-4d5a84cadf59c385da07effb81988b958c69a75743a5516c415216d5211b95c5 WatchSource:0}: Error finding container 4d5a84cadf59c385da07effb81988b958c69a75743a5516c415216d5211b95c5: Status 404 returned error can't find the container with id 4d5a84cadf59c385da07effb81988b958c69a75743a5516c415216d5211b95c5 Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.701902 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-pqd4c"] Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.719516 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-8wk6c"] Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.728246 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-cl4tl" event={"ID":"4d915797-cd52-4f07-9028-4f80411fe65b","Type":"ContainerStarted","Data":"ff4b2e27827df6668b31a96deb8e605684b173d8a481be00f4a61f3986054ed7"} Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.728450 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7875d8bb94-dn8xx"] Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.729582 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-8vknn" event={"ID":"69877edd-8b97-4838-bc41-c087fc4deb7e","Type":"ContainerStarted","Data":"7160808df148bce69e56df80b941a10a24e1b28f629f7733243c1d87c09fba28"} Nov 22 05:07:15 crc kubenswrapper[4818]: W1122 05:07:15.731955 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7389ff62_006d_4d11_8237_e57c9279b0f3.slice/crio-a34dcf761b3182659ad2d2228eabfe9ee7218aff13fa8ebaae73fbd4c3910490 WatchSource:0}: Error finding container a34dcf761b3182659ad2d2228eabfe9ee7218aff13fa8ebaae73fbd4c3910490: Status 404 returned error can't find the container with id a34dcf761b3182659ad2d2228eabfe9ee7218aff13fa8ebaae73fbd4c3910490 Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.732003 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4zl42" event={"ID":"05bab737-6cda-4891-a014-202ab05b141c","Type":"ContainerStarted","Data":"afd322bee71d6822fd5d1bb3b6d5199e46a1c8a02fca3b39e310e6a143881406"} Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.733701 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-qt9nt" event={"ID":"10da5c26-c5b3-4477-a4d4-a5c98741dfb1","Type":"ContainerStarted","Data":"008a7a152f9e416f70f9464c4bcb17b72bdbdb5dd128114587e05c657cf81164"} Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.734693 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-85bf8d6cb7-lzmmp" event={"ID":"4838407d-e782-4aab-8fbf-861f3ba9019a","Type":"ContainerStarted","Data":"4d5a84cadf59c385da07effb81988b958c69a75743a5516c415216d5211b95c5"} Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.736267 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-b2lp5"] Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.736704 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xtplb" event={"ID":"68668b7f-5fc3-4774-8f0d-3eae71839d52","Type":"ContainerStarted","Data":"1b952aa2f23ea03bc2ead25eb7a0deaf42a396929b19870b1e5bd207364bfc17"} Nov 22 05:07:15 crc kubenswrapper[4818]: I1122 05:07:15.738339 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-jttdm" event={"ID":"b9a85b40-dd76-45b2-bd41-ecb5be7387da","Type":"ContainerStarted","Data":"4f7b16eed80fe0891840c20db3694308113420d8500d37ad2162eae4cfab9c3f"} Nov 22 05:07:15 crc kubenswrapper[4818]: W1122 05:07:15.755343 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96a59329_8f96_43a5_9abc_fa8d1d7009fb.slice/crio-04c0c820a3ebd8338fb608057e0667f7e4e5fbe28be5cafb85d5219875f531e7 WatchSource:0}: Error finding container 04c0c820a3ebd8338fb608057e0667f7e4e5fbe28be5cafb85d5219875f531e7: Status 404 returned error can't find the container with id 04c0c820a3ebd8338fb608057e0667f7e4e5fbe28be5cafb85d5219875f531e7 Nov 22 05:07:15 crc kubenswrapper[4818]: W1122 05:07:15.757450 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27ec8b4f_e5ec_483a_ada8_6a766b843b73.slice/crio-8001018e46869e3766000cf7b27160fa42462f225e3c01af48d28d3d5495e39c WatchSource:0}: Error finding container 8001018e46869e3766000cf7b27160fa42462f225e3c01af48d28d3d5495e39c: Status 404 returned error can't find the container with id 8001018e46869e3766000cf7b27160fa42462f225e3c01af48d28d3d5495e39c Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.013565 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9797ea62-91f5-4b4f-a9f0-27911cf629e4-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-n54fq\" (UID: \"9797ea62-91f5-4b4f-a9f0-27911cf629e4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-n54fq" Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.037300 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9797ea62-91f5-4b4f-a9f0-27911cf629e4-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-n54fq\" (UID: \"9797ea62-91f5-4b4f-a9f0-27911cf629e4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-n54fq" Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.071964 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-sqvdq"] Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.112248 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d6957"] Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.119356 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-8dhvf"] Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.123071 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-vkj22"] Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.138091 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-n54fq" Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.139246 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-lx77w"] Nov 22 05:07:16 crc kubenswrapper[4818]: E1122 05:07:16.144706 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-g445z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-99b499f4-ncfxv_openstack-operators(c31258b3-6cac-41e8-a350-1ea7b360b848): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 22 05:07:16 crc kubenswrapper[4818]: W1122 05:07:16.148059 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod786ab73e_012b_47f6_b033_9fa9aab6d6cb.slice/crio-0610920bb4d37c8047d189d51dc6631c7597e72f335c2bd0e2787a345c52a70a WatchSource:0}: Error finding container 0610920bb4d37c8047d189d51dc6631c7597e72f335c2bd0e2787a345c52a70a: Status 404 returned error can't find the container with id 0610920bb4d37c8047d189d51dc6631c7597e72f335c2bd0e2787a345c52a70a Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.151387 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-cjphz"] Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.158928 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-ncfxv"] Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.163569 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-56phb"] Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.168038 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-rf9vq"] Nov 22 05:07:16 crc kubenswrapper[4818]: E1122 05:07:16.170977 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jqlnl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5b797b8dff-56phb_openstack-operators(786ab73e-012b-47f6-b033-9fa9aab6d6cb): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 22 05:07:16 crc kubenswrapper[4818]: E1122 05:07:16.171218 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vqhlq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-b4c496f69-pb45n_openstack-operators(e342ff02-d6d9-4be0-b97f-87dea6653a61): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.178071 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-pb45n"] Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.181475 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6b57b64d88-pnqbd"] Nov 22 05:07:16 crc kubenswrapper[4818]: E1122 05:07:16.182512 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-996gn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-d656998f4-vkj22_openstack-operators(f04a965b-f5f4-45af-843f-862498956853): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 22 05:07:16 crc kubenswrapper[4818]: W1122 05:07:16.202474 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10d7f6df_ae52_4dab_b49d_764d95f77935.slice/crio-c48ee552979813eeda7461ac7b6abfd8dbd69104746ab2aae1a089e367f2cc49 WatchSource:0}: Error finding container c48ee552979813eeda7461ac7b6abfd8dbd69104746ab2aae1a089e367f2cc49: Status 404 returned error can't find the container with id c48ee552979813eeda7461ac7b6abfd8dbd69104746ab2aae1a089e367f2cc49 Nov 22 05:07:16 crc kubenswrapper[4818]: E1122 05:07:16.224673 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dmxgx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-cfbb9c588-lx77w_openstack-operators(531548a2-b8d3-4a2a-a6e9-1bb2af1891d8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 22 05:07:16 crc kubenswrapper[4818]: E1122 05:07:16.404849 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-pb45n" podUID="e342ff02-d6d9-4be0-b97f-87dea6653a61" Nov 22 05:07:16 crc kubenswrapper[4818]: E1122 05:07:16.549367 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-vkj22" podUID="f04a965b-f5f4-45af-843f-862498956853" Nov 22 05:07:16 crc kubenswrapper[4818]: E1122 05:07:16.566567 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-lx77w" podUID="531548a2-b8d3-4a2a-a6e9-1bb2af1891d8" Nov 22 05:07:16 crc kubenswrapper[4818]: E1122 05:07:16.566938 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-56phb" podUID="786ab73e-012b-47f6-b033-9fa9aab6d6cb" Nov 22 05:07:16 crc kubenswrapper[4818]: E1122 05:07:16.591139 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ncfxv" podUID="c31258b3-6cac-41e8-a350-1ea7b360b848" Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.680214 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-n54fq"] Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.761655 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-dn8xx" event={"ID":"7389ff62-006d-4d11-8237-e57c9279b0f3","Type":"ContainerStarted","Data":"a34dcf761b3182659ad2d2228eabfe9ee7218aff13fa8ebaae73fbd4c3910490"} Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.774322 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-8dhvf" event={"ID":"06ec2990-cd45-4808-9748-2a1d75c5015f","Type":"ContainerStarted","Data":"3dbd5aa8d2f8d899fd7a5424af7470ff74d72e48d4331444737552039d0e11da"} Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.800431 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ncfxv" event={"ID":"c31258b3-6cac-41e8-a350-1ea7b360b848","Type":"ContainerStarted","Data":"7279aeb0a239785ea2c4d484a4c4caad9b110d0d55a15545e5b599d05ca3d1dc"} Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.800484 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ncfxv" event={"ID":"c31258b3-6cac-41e8-a350-1ea7b360b848","Type":"ContainerStarted","Data":"06f8367c1c06fa7f6fb9bcfaf0e48a18d9519e5f5f375fcd5e5cc8fd8f74a379"} Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.806559 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d6957" event={"ID":"62c4e1cf-72fb-4a9b-88a1-350c1e20f360","Type":"ContainerStarted","Data":"da12ba4cabb9270127350145af4260d37dbfaaa2a17c802286627c716c153f06"} Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.807875 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-cjphz" event={"ID":"bba055bd-511b-4af6-bc94-4cadf7a84853","Type":"ContainerStarted","Data":"c5f7b2ddd16a738091db18d7dfde937d7f13aa9affbbed5da0f66a0c0d6e798b"} Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.813637 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6b57b64d88-pnqbd" event={"ID":"44f3f34b-8fa2-479c-a876-947fab621708","Type":"ContainerStarted","Data":"ce934b5bc9e528ade0a6c0de6aded70cfb0b511dfdf55174f8a3914ad36238f5"} Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.813681 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6b57b64d88-pnqbd" event={"ID":"44f3f34b-8fa2-479c-a876-947fab621708","Type":"ContainerStarted","Data":"f2e06793cca15687d2bd7dbd36606457c8691d87cc414991a922be1e267dde5c"} Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.818296 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-pb45n" event={"ID":"e342ff02-d6d9-4be0-b97f-87dea6653a61","Type":"ContainerStarted","Data":"b94a0b2aca082ed866dbfaed65224d95bbe59ff68182f076b60b6693066cf68a"} Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.818368 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-pb45n" event={"ID":"e342ff02-d6d9-4be0-b97f-87dea6653a61","Type":"ContainerStarted","Data":"a4b80346ad8a541ab4b525010ed44def990a809b1feefd8e8098256568f2536f"} Nov 22 05:07:16 crc kubenswrapper[4818]: E1122 05:07:16.821747 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ncfxv" podUID="c31258b3-6cac-41e8-a350-1ea7b360b848" Nov 22 05:07:16 crc kubenswrapper[4818]: E1122 05:07:16.834809 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-pb45n" podUID="e342ff02-d6d9-4be0-b97f-87dea6653a61" Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.844522 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-rf9vq" event={"ID":"10d7f6df-ae52-4dab-b49d-764d95f77935","Type":"ContainerStarted","Data":"c48ee552979813eeda7461ac7b6abfd8dbd69104746ab2aae1a089e367f2cc49"} Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.846336 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-pqd4c" event={"ID":"d6b59804-efc5-4f60-b9d3-bf0e998c8f89","Type":"ContainerStarted","Data":"b98f8d1c66205af7034f3b4d44b632b48316156174b60aadd2c4001690abac4e"} Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.847318 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-8wk6c" event={"ID":"96a59329-8f96-43a5-9abc-fa8d1d7009fb","Type":"ContainerStarted","Data":"04c0c820a3ebd8338fb608057e0667f7e4e5fbe28be5cafb85d5219875f531e7"} Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.849618 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-vkj22" event={"ID":"f04a965b-f5f4-45af-843f-862498956853","Type":"ContainerStarted","Data":"f99a02b66acedf2b41dc15736e2e6bbe98fb862fa8f8427e9a93ae7e06a3d2dd"} Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.849660 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-vkj22" event={"ID":"f04a965b-f5f4-45af-843f-862498956853","Type":"ContainerStarted","Data":"60aa263a623acd26b3ad58045d30cce56138e9f17625d7a3cbfa9542f2541ebc"} Nov 22 05:07:16 crc kubenswrapper[4818]: E1122 05:07:16.902010 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-vkj22" podUID="f04a965b-f5f4-45af-843f-862498956853" Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.921536 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-b2lp5" event={"ID":"27ec8b4f-e5ec-483a-ada8-6a766b843b73","Type":"ContainerStarted","Data":"8001018e46869e3766000cf7b27160fa42462f225e3c01af48d28d3d5495e39c"} Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.950527 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-lx77w" event={"ID":"531548a2-b8d3-4a2a-a6e9-1bb2af1891d8","Type":"ContainerStarted","Data":"41ff3be518c02ab67cfd432534e985845befa36fbbcea3734c9f0254ad22e440"} Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.950571 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-lx77w" event={"ID":"531548a2-b8d3-4a2a-a6e9-1bb2af1891d8","Type":"ContainerStarted","Data":"510cb4d22a58fb0f4c090c9a4e6908e955b53518a1a70a1995a86af8c7560588"} Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.973076 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-sqvdq" event={"ID":"2b80bb62-ec21-44c6-91e2-b422b9813255","Type":"ContainerStarted","Data":"68d7fb013c75cf103dc039ad0909cb92b51c8a8a72981c15862dee688596bd77"} Nov 22 05:07:16 crc kubenswrapper[4818]: E1122 05:07:16.973211 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\"" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-lx77w" podUID="531548a2-b8d3-4a2a-a6e9-1bb2af1891d8" Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.974898 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-56phb" event={"ID":"786ab73e-012b-47f6-b033-9fa9aab6d6cb","Type":"ContainerStarted","Data":"bf59636b1f775fbe76459ac10eb4696eb6206b273a32ae9050b3c4e66313102f"} Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.974927 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-56phb" event={"ID":"786ab73e-012b-47f6-b033-9fa9aab6d6cb","Type":"ContainerStarted","Data":"0610920bb4d37c8047d189d51dc6631c7597e72f335c2bd0e2787a345c52a70a"} Nov 22 05:07:16 crc kubenswrapper[4818]: E1122 05:07:16.977027 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-56phb" podUID="786ab73e-012b-47f6-b033-9fa9aab6d6cb" Nov 22 05:07:16 crc kubenswrapper[4818]: I1122 05:07:16.977572 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-n54fq" event={"ID":"9797ea62-91f5-4b4f-a9f0-27911cf629e4","Type":"ContainerStarted","Data":"343f4f2abac967642f3ab36e435ef09c9c92a9ce32ba29875f6d4f05c8704292"} Nov 22 05:07:17 crc kubenswrapper[4818]: I1122 05:07:17.994641 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6b57b64d88-pnqbd" event={"ID":"44f3f34b-8fa2-479c-a876-947fab621708","Type":"ContainerStarted","Data":"349d74dc52074f5bdc926e6152f5e46a27c6ae32b7de4cb58e95ec1af6601ac4"} Nov 22 05:07:17 crc kubenswrapper[4818]: I1122 05:07:17.994961 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6b57b64d88-pnqbd" Nov 22 05:07:17 crc kubenswrapper[4818]: E1122 05:07:17.997601 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-56phb" podUID="786ab73e-012b-47f6-b033-9fa9aab6d6cb" Nov 22 05:07:17 crc kubenswrapper[4818]: E1122 05:07:17.998594 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-pb45n" podUID="e342ff02-d6d9-4be0-b97f-87dea6653a61" Nov 22 05:07:17 crc kubenswrapper[4818]: E1122 05:07:17.999313 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\"" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-lx77w" podUID="531548a2-b8d3-4a2a-a6e9-1bb2af1891d8" Nov 22 05:07:18 crc kubenswrapper[4818]: E1122 05:07:18.025068 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-vkj22" podUID="f04a965b-f5f4-45af-843f-862498956853" Nov 22 05:07:18 crc kubenswrapper[4818]: E1122 05:07:18.025162 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ncfxv" podUID="c31258b3-6cac-41e8-a350-1ea7b360b848" Nov 22 05:07:18 crc kubenswrapper[4818]: I1122 05:07:18.052583 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6b57b64d88-pnqbd" podStartSLOduration=4.052548987 podStartE2EDuration="4.052548987s" podCreationTimestamp="2025-11-22 05:07:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:07:18.045959712 +0000 UTC m=+1190.620376259" watchObservedRunningTime="2025-11-22 05:07:18.052548987 +0000 UTC m=+1190.626965514" Nov 22 05:07:21 crc kubenswrapper[4818]: I1122 05:07:21.264614 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:07:21 crc kubenswrapper[4818]: I1122 05:07:21.264954 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:07:25 crc kubenswrapper[4818]: I1122 05:07:25.472715 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6b57b64d88-pnqbd" Nov 22 05:07:28 crc kubenswrapper[4818]: E1122 05:07:28.400416 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:848f4c43c6bdd4e33e3ce1d147a85b9b6a6124a150bd5155dce421ef539259e9" Nov 22 05:07:28 crc kubenswrapper[4818]: E1122 05:07:28.400972 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:848f4c43c6bdd4e33e3ce1d147a85b9b6a6124a150bd5155dce421ef539259e9,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w4ggr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-598f69df5d-4zl42_openstack-operators(05bab737-6cda-4891-a014-202ab05b141c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 05:07:31 crc kubenswrapper[4818]: E1122 05:07:31.016008 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f" Nov 22 05:07:31 crc kubenswrapper[4818]: E1122 05:07:31.016556 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-66zvd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-8c6448b9f-sqvdq_openstack-operators(2b80bb62-ec21-44c6-91e2-b422b9813255): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 05:07:31 crc kubenswrapper[4818]: I1122 05:07:31.429081 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 05:07:31 crc kubenswrapper[4818]: E1122 05:07:31.649119 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.5:5001/openstack-k8s-operators/manila-operator:29c64873c2770ee4f517c4c15766d47d511eceb0" Nov 22 05:07:31 crc kubenswrapper[4818]: E1122 05:07:31.649452 4818 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.5:5001/openstack-k8s-operators/manila-operator:29c64873c2770ee4f517c4c15766d47d511eceb0" Nov 22 05:07:31 crc kubenswrapper[4818]: E1122 05:07:31.649625 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.5:5001/openstack-k8s-operators/manila-operator:29c64873c2770ee4f517c4c15766d47d511eceb0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wxwc5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-85bf8d6cb7-lzmmp_openstack-operators(4838407d-e782-4aab-8fbf-861f3ba9019a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 05:07:32 crc kubenswrapper[4818]: E1122 05:07:32.255891 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Nov 22 05:07:32 crc kubenswrapper[4818]: E1122 05:07:32.256157 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xr9dc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-d6957_openstack-operators(62c4e1cf-72fb-4a9b-88a1-350c1e20f360): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 05:07:32 crc kubenswrapper[4818]: E1122 05:07:32.257496 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d6957" podUID="62c4e1cf-72fb-4a9b-88a1-350c1e20f360" Nov 22 05:07:33 crc kubenswrapper[4818]: E1122 05:07:33.120663 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d6957" podUID="62c4e1cf-72fb-4a9b-88a1-350c1e20f360" Nov 22 05:07:36 crc kubenswrapper[4818]: I1122 05:07:36.142364 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-cl4tl" event={"ID":"4d915797-cd52-4f07-9028-4f80411fe65b","Type":"ContainerStarted","Data":"8669bf84ba4af43e5b5ec3f6c653d1703d12344dc805348b2178178e290f7639"} Nov 22 05:07:36 crc kubenswrapper[4818]: I1122 05:07:36.144477 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4zl42" event={"ID":"05bab737-6cda-4891-a014-202ab05b141c","Type":"ContainerStarted","Data":"003289f592a8435144c2d1231a6751025d0f061fe5c698c41ad20a9c050d28e6"} Nov 22 05:07:36 crc kubenswrapper[4818]: I1122 05:07:36.145931 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-pqd4c" event={"ID":"d6b59804-efc5-4f60-b9d3-bf0e998c8f89","Type":"ContainerStarted","Data":"8df247a38ea96b6edac710e071a748388884675231fc9b6b95a2ab8a7097fcb8"} Nov 22 05:07:36 crc kubenswrapper[4818]: I1122 05:07:36.147243 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-rf9vq" event={"ID":"10d7f6df-ae52-4dab-b49d-764d95f77935","Type":"ContainerStarted","Data":"2d56de77a4dd1a465a65b651911ff4b71aacb02edb34531e2b69f359814f3f55"} Nov 22 05:07:36 crc kubenswrapper[4818]: I1122 05:07:36.148420 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-dn8xx" event={"ID":"7389ff62-006d-4d11-8237-e57c9279b0f3","Type":"ContainerStarted","Data":"32fe696783d744f64346d2b7d83971ebdb9e0189d6b4b7f2949c103804e5e479"} Nov 22 05:07:36 crc kubenswrapper[4818]: I1122 05:07:36.150165 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-b2lp5" event={"ID":"27ec8b4f-e5ec-483a-ada8-6a766b843b73","Type":"ContainerStarted","Data":"52b7329889493de019b42e29676701b19b488139394ad044e4fd7844ba60dff6"} Nov 22 05:07:36 crc kubenswrapper[4818]: I1122 05:07:36.151371 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-8dhvf" event={"ID":"06ec2990-cd45-4808-9748-2a1d75c5015f","Type":"ContainerStarted","Data":"a80f7792dc1feb1e1fee547b406bdaa8eaf7ea1c28c95b00f378c1151e4931bb"} Nov 22 05:07:36 crc kubenswrapper[4818]: I1122 05:07:36.152507 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-qt9nt" event={"ID":"10da5c26-c5b3-4477-a4d4-a5c98741dfb1","Type":"ContainerStarted","Data":"e88c435cef844cae38798c2756d4706d3cd1cd5e10bdbb8549601a36e2814d42"} Nov 22 05:07:36 crc kubenswrapper[4818]: I1122 05:07:36.161777 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-85bf8d6cb7-lzmmp" event={"ID":"4838407d-e782-4aab-8fbf-861f3ba9019a","Type":"ContainerStarted","Data":"fd832e91e7a368ea0eea590cb8ea3e4b9eeca5f264f35ce1bc44cf9f01e5f229"} Nov 22 05:07:36 crc kubenswrapper[4818]: I1122 05:07:36.163176 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-n54fq" event={"ID":"9797ea62-91f5-4b4f-a9f0-27911cf629e4","Type":"ContainerStarted","Data":"9f018b56d93b9a91fa5d96a7f585c0267036597c564cefd3655727be1a4a3bf6"} Nov 22 05:07:36 crc kubenswrapper[4818]: I1122 05:07:36.164408 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-8vknn" event={"ID":"69877edd-8b97-4838-bc41-c087fc4deb7e","Type":"ContainerStarted","Data":"f6fc006986aa1625001e68826eae717ed13979c7248a690f4a77eabee1042eee"} Nov 22 05:07:36 crc kubenswrapper[4818]: I1122 05:07:36.167151 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-jttdm" event={"ID":"b9a85b40-dd76-45b2-bd41-ecb5be7387da","Type":"ContainerStarted","Data":"59b55c6f8f23c93a70eb5d12ba7d5de2d5e8dc484cdb64bc6cb4e92addab1a8f"} Nov 22 05:07:36 crc kubenswrapper[4818]: I1122 05:07:36.168416 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-sqvdq" event={"ID":"2b80bb62-ec21-44c6-91e2-b422b9813255","Type":"ContainerStarted","Data":"0b72d15ba1f226adb1ec4bbfeb6e4f5444f2313e982e1ab44b623fd5e73205b9"} Nov 22 05:07:36 crc kubenswrapper[4818]: I1122 05:07:36.169584 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xtplb" event={"ID":"68668b7f-5fc3-4774-8f0d-3eae71839d52","Type":"ContainerStarted","Data":"6a9918b99ad3732cdb023c9271ebc9325f071cff872d3d74522e0119b9cf9235"} Nov 22 05:07:36 crc kubenswrapper[4818]: I1122 05:07:36.170777 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-cjphz" event={"ID":"bba055bd-511b-4af6-bc94-4cadf7a84853","Type":"ContainerStarted","Data":"cb9202b14a489afd0fdfc7556a734d66c3d2f2d6b4b4a37b406cced0c134d6ee"} Nov 22 05:07:36 crc kubenswrapper[4818]: I1122 05:07:36.172075 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-8wk6c" event={"ID":"96a59329-8f96-43a5-9abc-fa8d1d7009fb","Type":"ContainerStarted","Data":"779e4dbb1e116a09d09266d4fa216311232524d8e8ec645234418f1efb19c11e"} Nov 22 05:07:37 crc kubenswrapper[4818]: E1122 05:07:37.123120 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-85bf8d6cb7-lzmmp" podUID="4838407d-e782-4aab-8fbf-861f3ba9019a" Nov 22 05:07:37 crc kubenswrapper[4818]: E1122 05:07:37.123272 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4zl42" podUID="05bab737-6cda-4891-a014-202ab05b141c" Nov 22 05:07:37 crc kubenswrapper[4818]: E1122 05:07:37.123597 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-sqvdq" podUID="2b80bb62-ec21-44c6-91e2-b422b9813255" Nov 22 05:07:37 crc kubenswrapper[4818]: I1122 05:07:37.186941 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-jttdm" event={"ID":"b9a85b40-dd76-45b2-bd41-ecb5be7387da","Type":"ContainerStarted","Data":"e45f24a8df13a4b35a03fa15f6800520b68659ca28617a38946e74930c5efd88"} Nov 22 05:07:37 crc kubenswrapper[4818]: E1122 05:07:37.194155 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:848f4c43c6bdd4e33e3ce1d147a85b9b6a6124a150bd5155dce421ef539259e9\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4zl42" podUID="05bab737-6cda-4891-a014-202ab05b141c" Nov 22 05:07:37 crc kubenswrapper[4818]: E1122 05:07:37.194272 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-sqvdq" podUID="2b80bb62-ec21-44c6-91e2-b422b9813255" Nov 22 05:07:37 crc kubenswrapper[4818]: E1122 05:07:37.194737 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.5:5001/openstack-k8s-operators/manila-operator:29c64873c2770ee4f517c4c15766d47d511eceb0\\\"\"" pod="openstack-operators/manila-operator-controller-manager-85bf8d6cb7-lzmmp" podUID="4838407d-e782-4aab-8fbf-861f3ba9019a" Nov 22 05:07:40 crc kubenswrapper[4818]: I1122 05:07:40.216087 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-b2lp5" event={"ID":"27ec8b4f-e5ec-483a-ada8-6a766b843b73","Type":"ContainerStarted","Data":"c13e101fc7cc0cdfa9cd7b7ad331218b19bf2288119137660aa1f4dd4c18b73a"} Nov 22 05:07:40 crc kubenswrapper[4818]: I1122 05:07:40.218200 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xtplb" event={"ID":"68668b7f-5fc3-4774-8f0d-3eae71839d52","Type":"ContainerStarted","Data":"6095625b2a75e2fa0c5cbcac15ae32669709ae0419ecb22b247e7e0af09a07da"} Nov 22 05:07:40 crc kubenswrapper[4818]: I1122 05:07:40.220128 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-cjphz" event={"ID":"bba055bd-511b-4af6-bc94-4cadf7a84853","Type":"ContainerStarted","Data":"c0b44526b0ad9fc1b4b3f49d64cb2f6652eeef8b1cf517b9893a778ce2584600"} Nov 22 05:07:41 crc kubenswrapper[4818]: I1122 05:07:41.231211 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-qt9nt" event={"ID":"10da5c26-c5b3-4477-a4d4-a5c98741dfb1","Type":"ContainerStarted","Data":"bfd6195689006fffd170c4617945e6fde4c37188fa1753e5ec8a510cb47decef"} Nov 22 05:07:41 crc kubenswrapper[4818]: I1122 05:07:41.233370 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-pqd4c" event={"ID":"d6b59804-efc5-4f60-b9d3-bf0e998c8f89","Type":"ContainerStarted","Data":"74bc16fc51e1ae06df0cb2b731f1e0d6e58223650a8b415aac56bc3b80fb93bc"} Nov 22 05:07:41 crc kubenswrapper[4818]: I1122 05:07:41.236451 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-n54fq" event={"ID":"9797ea62-91f5-4b4f-a9f0-27911cf629e4","Type":"ContainerStarted","Data":"b4038c6e9941732a5986d6d50e41886e9bcbec1f40d657e18613f4db6cb34e27"} Nov 22 05:07:41 crc kubenswrapper[4818]: I1122 05:07:41.239064 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-rf9vq" event={"ID":"10d7f6df-ae52-4dab-b49d-764d95f77935","Type":"ContainerStarted","Data":"597a50fe8a95af8bd8937419d41b97ff23548d3ff49da854519beb47df78277c"} Nov 22 05:07:41 crc kubenswrapper[4818]: I1122 05:07:41.241474 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-dn8xx" event={"ID":"7389ff62-006d-4d11-8237-e57c9279b0f3","Type":"ContainerStarted","Data":"fc3562d888a3d9be1d56ac692ea13765b2d07716ad8fa44caa4945a74ab49166"} Nov 22 05:07:41 crc kubenswrapper[4818]: I1122 05:07:41.243740 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-8wk6c" event={"ID":"96a59329-8f96-43a5-9abc-fa8d1d7009fb","Type":"ContainerStarted","Data":"7b43704b6da6bcffda75ed68966b499440ab906542aad0dc0648e9fe34de2b15"} Nov 22 05:07:41 crc kubenswrapper[4818]: I1122 05:07:41.245803 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-8dhvf" event={"ID":"06ec2990-cd45-4808-9748-2a1d75c5015f","Type":"ContainerStarted","Data":"258c9104d1bdfdec9d541c1493a4c0a8244750d11526262d9ce43413bbc0b811"} Nov 22 05:07:41 crc kubenswrapper[4818]: I1122 05:07:41.248178 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-cl4tl" event={"ID":"4d915797-cd52-4f07-9028-4f80411fe65b","Type":"ContainerStarted","Data":"de2d841b0c9820feef93b921ce70205688005e5aed04fe7285625266ab9b68f5"} Nov 22 05:07:41 crc kubenswrapper[4818]: I1122 05:07:41.252844 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-8vknn" event={"ID":"69877edd-8b97-4838-bc41-c087fc4deb7e","Type":"ContainerStarted","Data":"b31adbf3afe19ebc71168bbf262571b9fc4db502b2b16be427406ac3206e7ae8"} Nov 22 05:07:41 crc kubenswrapper[4818]: I1122 05:07:41.253062 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-cjphz" Nov 22 05:07:41 crc kubenswrapper[4818]: I1122 05:07:41.256989 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-cjphz" Nov 22 05:07:41 crc kubenswrapper[4818]: I1122 05:07:41.282307 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-cjphz" podStartSLOduration=11.320220013 podStartE2EDuration="27.282287657s" podCreationTimestamp="2025-11-22 05:07:14 +0000 UTC" firstStartedPulling="2025-11-22 05:07:16.139423978 +0000 UTC m=+1188.713840505" lastFinishedPulling="2025-11-22 05:07:32.101491582 +0000 UTC m=+1204.675908149" observedRunningTime="2025-11-22 05:07:41.277007647 +0000 UTC m=+1213.851424194" watchObservedRunningTime="2025-11-22 05:07:41.282287657 +0000 UTC m=+1213.856704204" Nov 22 05:07:41 crc kubenswrapper[4818]: I1122 05:07:41.310338 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-jttdm" podStartSLOduration=11.988498588 podStartE2EDuration="28.309882293s" podCreationTimestamp="2025-11-22 05:07:13 +0000 UTC" firstStartedPulling="2025-11-22 05:07:15.023571596 +0000 UTC m=+1187.597988123" lastFinishedPulling="2025-11-22 05:07:31.344955261 +0000 UTC m=+1203.919371828" observedRunningTime="2025-11-22 05:07:41.300173704 +0000 UTC m=+1213.874590241" watchObservedRunningTime="2025-11-22 05:07:41.309882293 +0000 UTC m=+1213.884298860" Nov 22 05:07:41 crc kubenswrapper[4818]: I1122 05:07:41.361562 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xtplb" podStartSLOduration=11.851848564 podStartE2EDuration="28.36154244s" podCreationTimestamp="2025-11-22 05:07:13 +0000 UTC" firstStartedPulling="2025-11-22 05:07:15.587874162 +0000 UTC m=+1188.162290689" lastFinishedPulling="2025-11-22 05:07:32.097568038 +0000 UTC m=+1204.671984565" observedRunningTime="2025-11-22 05:07:41.35517387 +0000 UTC m=+1213.929590407" watchObservedRunningTime="2025-11-22 05:07:41.36154244 +0000 UTC m=+1213.935958967" Nov 22 05:07:42 crc kubenswrapper[4818]: I1122 05:07:42.269967 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-rf9vq" Nov 22 05:07:42 crc kubenswrapper[4818]: I1122 05:07:42.271342 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-qt9nt" Nov 22 05:07:42 crc kubenswrapper[4818]: I1122 05:07:42.271763 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-8dhvf" Nov 22 05:07:42 crc kubenswrapper[4818]: I1122 05:07:42.273564 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-rf9vq" Nov 22 05:07:42 crc kubenswrapper[4818]: I1122 05:07:42.273856 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-qt9nt" Nov 22 05:07:42 crc kubenswrapper[4818]: I1122 05:07:42.274142 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-8dhvf" Nov 22 05:07:42 crc kubenswrapper[4818]: I1122 05:07:42.292290 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-qt9nt" podStartSLOduration=12.802348628 podStartE2EDuration="29.292230255s" podCreationTimestamp="2025-11-22 05:07:13 +0000 UTC" firstStartedPulling="2025-11-22 05:07:15.607730842 +0000 UTC m=+1188.182147379" lastFinishedPulling="2025-11-22 05:07:32.097612439 +0000 UTC m=+1204.672029006" observedRunningTime="2025-11-22 05:07:42.288349472 +0000 UTC m=+1214.862765999" watchObservedRunningTime="2025-11-22 05:07:42.292230255 +0000 UTC m=+1214.866646782" Nov 22 05:07:42 crc kubenswrapper[4818]: I1122 05:07:42.320205 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-n54fq" podStartSLOduration=12.960260212 podStartE2EDuration="28.320186971s" podCreationTimestamp="2025-11-22 05:07:14 +0000 UTC" firstStartedPulling="2025-11-22 05:07:16.741651405 +0000 UTC m=+1189.316067932" lastFinishedPulling="2025-11-22 05:07:32.101578124 +0000 UTC m=+1204.675994691" observedRunningTime="2025-11-22 05:07:42.316943344 +0000 UTC m=+1214.891359871" watchObservedRunningTime="2025-11-22 05:07:42.320186971 +0000 UTC m=+1214.894603498" Nov 22 05:07:42 crc kubenswrapper[4818]: I1122 05:07:42.349597 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-dn8xx" podStartSLOduration=12.990067433 podStartE2EDuration="29.349574774s" podCreationTimestamp="2025-11-22 05:07:13 +0000 UTC" firstStartedPulling="2025-11-22 05:07:15.734994265 +0000 UTC m=+1188.309410802" lastFinishedPulling="2025-11-22 05:07:32.094501576 +0000 UTC m=+1204.668918143" observedRunningTime="2025-11-22 05:07:42.343751389 +0000 UTC m=+1214.918167916" watchObservedRunningTime="2025-11-22 05:07:42.349574774 +0000 UTC m=+1214.923991301" Nov 22 05:07:42 crc kubenswrapper[4818]: I1122 05:07:42.369085 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-rf9vq" podStartSLOduration=13.495070147 podStartE2EDuration="29.369064044s" podCreationTimestamp="2025-11-22 05:07:13 +0000 UTC" firstStartedPulling="2025-11-22 05:07:16.223821878 +0000 UTC m=+1188.798238405" lastFinishedPulling="2025-11-22 05:07:32.097815735 +0000 UTC m=+1204.672232302" observedRunningTime="2025-11-22 05:07:42.361764129 +0000 UTC m=+1214.936180656" watchObservedRunningTime="2025-11-22 05:07:42.369064044 +0000 UTC m=+1214.943480571" Nov 22 05:07:42 crc kubenswrapper[4818]: I1122 05:07:42.388745 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-8dhvf" podStartSLOduration=12.420636534 podStartE2EDuration="28.388716988s" podCreationTimestamp="2025-11-22 05:07:14 +0000 UTC" firstStartedPulling="2025-11-22 05:07:16.126947786 +0000 UTC m=+1188.701364313" lastFinishedPulling="2025-11-22 05:07:32.09502824 +0000 UTC m=+1204.669444767" observedRunningTime="2025-11-22 05:07:42.379850381 +0000 UTC m=+1214.954266918" watchObservedRunningTime="2025-11-22 05:07:42.388716988 +0000 UTC m=+1214.963133555" Nov 22 05:07:42 crc kubenswrapper[4818]: I1122 05:07:42.402053 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-b2lp5" podStartSLOduration=13.821840161 podStartE2EDuration="29.402033153s" podCreationTimestamp="2025-11-22 05:07:13 +0000 UTC" firstStartedPulling="2025-11-22 05:07:15.763910876 +0000 UTC m=+1188.338327403" lastFinishedPulling="2025-11-22 05:07:31.344103828 +0000 UTC m=+1203.918520395" observedRunningTime="2025-11-22 05:07:42.396817933 +0000 UTC m=+1214.971234500" watchObservedRunningTime="2025-11-22 05:07:42.402033153 +0000 UTC m=+1214.976449680" Nov 22 05:07:42 crc kubenswrapper[4818]: I1122 05:07:42.421040 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-8wk6c" podStartSLOduration=13.079467006 podStartE2EDuration="29.421015609s" podCreationTimestamp="2025-11-22 05:07:13 +0000 UTC" firstStartedPulling="2025-11-22 05:07:15.757763022 +0000 UTC m=+1188.332179549" lastFinishedPulling="2025-11-22 05:07:32.099311585 +0000 UTC m=+1204.673728152" observedRunningTime="2025-11-22 05:07:42.412454751 +0000 UTC m=+1214.986871288" watchObservedRunningTime="2025-11-22 05:07:42.421015609 +0000 UTC m=+1214.995432136" Nov 22 05:07:42 crc kubenswrapper[4818]: I1122 05:07:42.465377 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7969689c84-cl4tl" podStartSLOduration=12.35538785 podStartE2EDuration="29.465351501s" podCreationTimestamp="2025-11-22 05:07:13 +0000 UTC" firstStartedPulling="2025-11-22 05:07:15.003650485 +0000 UTC m=+1187.578067002" lastFinishedPulling="2025-11-22 05:07:32.113614126 +0000 UTC m=+1204.688030653" observedRunningTime="2025-11-22 05:07:42.463764978 +0000 UTC m=+1215.038181515" watchObservedRunningTime="2025-11-22 05:07:42.465351501 +0000 UTC m=+1215.039768048" Nov 22 05:07:42 crc kubenswrapper[4818]: I1122 05:07:42.491917 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-pqd4c" podStartSLOduration=13.097826975 podStartE2EDuration="29.491895948s" podCreationTimestamp="2025-11-22 05:07:13 +0000 UTC" firstStartedPulling="2025-11-22 05:07:15.716785759 +0000 UTC m=+1188.291202286" lastFinishedPulling="2025-11-22 05:07:32.110854732 +0000 UTC m=+1204.685271259" observedRunningTime="2025-11-22 05:07:42.487942834 +0000 UTC m=+1215.062359371" watchObservedRunningTime="2025-11-22 05:07:42.491895948 +0000 UTC m=+1215.066312485" Nov 22 05:07:42 crc kubenswrapper[4818]: I1122 05:07:42.557112 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-8vknn" podStartSLOduration=12.238034362 podStartE2EDuration="29.557082527s" podCreationTimestamp="2025-11-22 05:07:13 +0000 UTC" firstStartedPulling="2025-11-22 05:07:14.772461021 +0000 UTC m=+1187.346877548" lastFinishedPulling="2025-11-22 05:07:32.091509186 +0000 UTC m=+1204.665925713" observedRunningTime="2025-11-22 05:07:42.550637685 +0000 UTC m=+1215.125054222" watchObservedRunningTime="2025-11-22 05:07:42.557082527 +0000 UTC m=+1215.131499094" Nov 22 05:07:44 crc kubenswrapper[4818]: I1122 05:07:44.074702 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-8vknn" Nov 22 05:07:44 crc kubenswrapper[4818]: I1122 05:07:44.077297 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-8vknn" Nov 22 05:07:44 crc kubenswrapper[4818]: I1122 05:07:44.175547 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-jttdm" Nov 22 05:07:44 crc kubenswrapper[4818]: I1122 05:07:44.178710 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-jttdm" Nov 22 05:07:44 crc kubenswrapper[4818]: I1122 05:07:44.188892 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7969689c84-cl4tl" Nov 22 05:07:44 crc kubenswrapper[4818]: I1122 05:07:44.190764 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7969689c84-cl4tl" Nov 22 05:07:44 crc kubenswrapper[4818]: I1122 05:07:44.395415 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-b2lp5" Nov 22 05:07:44 crc kubenswrapper[4818]: I1122 05:07:44.395618 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-pqd4c" Nov 22 05:07:44 crc kubenswrapper[4818]: I1122 05:07:44.397032 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-pqd4c" Nov 22 05:07:44 crc kubenswrapper[4818]: I1122 05:07:44.400611 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-b2lp5" Nov 22 05:07:44 crc kubenswrapper[4818]: I1122 05:07:44.563827 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-8wk6c" Nov 22 05:07:44 crc kubenswrapper[4818]: I1122 05:07:44.566365 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-8wk6c" Nov 22 05:07:44 crc kubenswrapper[4818]: I1122 05:07:44.586446 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xtplb" Nov 22 05:07:44 crc kubenswrapper[4818]: I1122 05:07:44.589411 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xtplb" Nov 22 05:07:44 crc kubenswrapper[4818]: I1122 05:07:44.935146 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-dn8xx" Nov 22 05:07:44 crc kubenswrapper[4818]: I1122 05:07:44.940849 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-dn8xx" Nov 22 05:07:46 crc kubenswrapper[4818]: I1122 05:07:46.141602 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-n54fq" Nov 22 05:07:46 crc kubenswrapper[4818]: I1122 05:07:46.152874 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-n54fq" Nov 22 05:07:51 crc kubenswrapper[4818]: I1122 05:07:51.265529 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:07:51 crc kubenswrapper[4818]: I1122 05:07:51.266441 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:07:53 crc kubenswrapper[4818]: E1122 05:07:53.393566 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7" Nov 22 05:07:53 crc kubenswrapper[4818]: E1122 05:07:53.393890 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dmxgx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-cfbb9c588-lx77w_openstack-operators(531548a2-b8d3-4a2a-a6e9-1bb2af1891d8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 05:07:53 crc kubenswrapper[4818]: E1122 05:07:53.395181 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-lx77w" podUID="531548a2-b8d3-4a2a-a6e9-1bb2af1891d8" Nov 22 05:07:58 crc kubenswrapper[4818]: E1122 05:07:58.707225 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c" Nov 22 05:07:58 crc kubenswrapper[4818]: E1122 05:07:58.707994 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jqlnl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5b797b8dff-56phb_openstack-operators(786ab73e-012b-47f6-b033-9fa9aab6d6cb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 05:07:58 crc kubenswrapper[4818]: E1122 05:07:58.709336 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-56phb" podUID="786ab73e-012b-47f6-b033-9fa9aab6d6cb" Nov 22 05:07:59 crc kubenswrapper[4818]: E1122 05:07:59.251933 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d" Nov 22 05:07:59 crc kubenswrapper[4818]: E1122 05:07:59.252181 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vqhlq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-b4c496f69-pb45n_openstack-operators(e342ff02-d6d9-4be0-b97f-87dea6653a61): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 05:07:59 crc kubenswrapper[4818]: E1122 05:07:59.253408 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-pb45n" podUID="e342ff02-d6d9-4be0-b97f-87dea6653a61" Nov 22 05:07:59 crc kubenswrapper[4818]: E1122 05:07:59.692743 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Nov 22 05:07:59 crc kubenswrapper[4818]: E1122 05:07:59.692906 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xr9dc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-d6957_openstack-operators(62c4e1cf-72fb-4a9b-88a1-350c1e20f360): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 05:07:59 crc kubenswrapper[4818]: E1122 05:07:59.694082 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d6957" podUID="62c4e1cf-72fb-4a9b-88a1-350c1e20f360" Nov 22 05:07:59 crc kubenswrapper[4818]: E1122 05:07:59.735631 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0" Nov 22 05:07:59 crc kubenswrapper[4818]: E1122 05:07:59.735844 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-996gn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-d656998f4-vkj22_openstack-operators(f04a965b-f5f4-45af-843f-862498956853): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 05:07:59 crc kubenswrapper[4818]: E1122 05:07:59.737030 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-vkj22" podUID="f04a965b-f5f4-45af-843f-862498956853" Nov 22 05:08:00 crc kubenswrapper[4818]: E1122 05:08:00.250399 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377" Nov 22 05:08:00 crc kubenswrapper[4818]: E1122 05:08:00.250873 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-g445z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-99b499f4-ncfxv_openstack-operators(c31258b3-6cac-41e8-a350-1ea7b360b848): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 05:08:00 crc kubenswrapper[4818]: E1122 05:08:00.252471 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ncfxv" podUID="c31258b3-6cac-41e8-a350-1ea7b360b848" Nov 22 05:08:01 crc kubenswrapper[4818]: I1122 05:08:01.434182 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4zl42" event={"ID":"05bab737-6cda-4891-a014-202ab05b141c","Type":"ContainerStarted","Data":"358bd17fe5f1db8074ed5e91427a8f21d1eaff95b48cc77cbeccb0a612027403"} Nov 22 05:08:01 crc kubenswrapper[4818]: I1122 05:08:01.435311 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4zl42" Nov 22 05:08:01 crc kubenswrapper[4818]: I1122 05:08:01.436923 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-85bf8d6cb7-lzmmp" event={"ID":"4838407d-e782-4aab-8fbf-861f3ba9019a","Type":"ContainerStarted","Data":"47e66dc221385db96a844811d30147c15b33b07c583869626174cf72df62d315"} Nov 22 05:08:01 crc kubenswrapper[4818]: I1122 05:08:01.437185 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-85bf8d6cb7-lzmmp" Nov 22 05:08:01 crc kubenswrapper[4818]: I1122 05:08:01.442873 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-sqvdq" event={"ID":"2b80bb62-ec21-44c6-91e2-b422b9813255","Type":"ContainerStarted","Data":"9cd65896a0c37163c15bae1a46cadca90c66ccf9b9402ea0b0358c566b9d853f"} Nov 22 05:08:01 crc kubenswrapper[4818]: I1122 05:08:01.443134 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-sqvdq" Nov 22 05:08:01 crc kubenswrapper[4818]: I1122 05:08:01.473723 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4zl42" podStartSLOduration=3.670928928 podStartE2EDuration="48.473687175s" podCreationTimestamp="2025-11-22 05:07:13 +0000 UTC" firstStartedPulling="2025-11-22 05:07:15.587446741 +0000 UTC m=+1188.161863268" lastFinishedPulling="2025-11-22 05:08:00.390204978 +0000 UTC m=+1232.964621515" observedRunningTime="2025-11-22 05:08:01.465124367 +0000 UTC m=+1234.039540924" watchObservedRunningTime="2025-11-22 05:08:01.473687175 +0000 UTC m=+1234.048103752" Nov 22 05:08:01 crc kubenswrapper[4818]: I1122 05:08:01.496123 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-85bf8d6cb7-lzmmp" podStartSLOduration=3.809740491 podStartE2EDuration="48.496103583s" podCreationTimestamp="2025-11-22 05:07:13 +0000 UTC" firstStartedPulling="2025-11-22 05:07:15.705990042 +0000 UTC m=+1188.280406559" lastFinishedPulling="2025-11-22 05:08:00.392353124 +0000 UTC m=+1232.966769651" observedRunningTime="2025-11-22 05:08:01.488198843 +0000 UTC m=+1234.062615390" watchObservedRunningTime="2025-11-22 05:08:01.496103583 +0000 UTC m=+1234.070520120" Nov 22 05:08:01 crc kubenswrapper[4818]: I1122 05:08:01.511759 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-sqvdq" podStartSLOduration=3.241664278 podStartE2EDuration="47.511737961s" podCreationTimestamp="2025-11-22 05:07:14 +0000 UTC" firstStartedPulling="2025-11-22 05:07:16.120813572 +0000 UTC m=+1188.695230099" lastFinishedPulling="2025-11-22 05:08:00.390887245 +0000 UTC m=+1232.965303782" observedRunningTime="2025-11-22 05:08:01.508660878 +0000 UTC m=+1234.083077445" watchObservedRunningTime="2025-11-22 05:08:01.511737961 +0000 UTC m=+1234.086154498" Nov 22 05:08:05 crc kubenswrapper[4818]: I1122 05:08:05.106630 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-sqvdq" Nov 22 05:08:07 crc kubenswrapper[4818]: E1122 05:08:07.294163 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\"" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-lx77w" podUID="531548a2-b8d3-4a2a-a6e9-1bb2af1891d8" Nov 22 05:08:10 crc kubenswrapper[4818]: E1122 05:08:10.294332 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d6957" podUID="62c4e1cf-72fb-4a9b-88a1-350c1e20f360" Nov 22 05:08:11 crc kubenswrapper[4818]: E1122 05:08:11.293489 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-56phb" podUID="786ab73e-012b-47f6-b033-9fa9aab6d6cb" Nov 22 05:08:12 crc kubenswrapper[4818]: E1122 05:08:12.292786 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ncfxv" podUID="c31258b3-6cac-41e8-a350-1ea7b360b848" Nov 22 05:08:13 crc kubenswrapper[4818]: E1122 05:08:13.294398 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-pb45n" podUID="e342ff02-d6d9-4be0-b97f-87dea6653a61" Nov 22 05:08:14 crc kubenswrapper[4818]: I1122 05:08:14.380437 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-4zl42" Nov 22 05:08:14 crc kubenswrapper[4818]: I1122 05:08:14.555762 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-85bf8d6cb7-lzmmp" Nov 22 05:08:15 crc kubenswrapper[4818]: E1122 05:08:15.292004 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-vkj22" podUID="f04a965b-f5f4-45af-843f-862498956853" Nov 22 05:08:21 crc kubenswrapper[4818]: I1122 05:08:21.265017 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:08:21 crc kubenswrapper[4818]: I1122 05:08:21.265603 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:08:21 crc kubenswrapper[4818]: I1122 05:08:21.265659 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 05:08:21 crc kubenswrapper[4818]: I1122 05:08:21.266379 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7f4e621532f88c0da5dac82557bd8694f552742e853952ce47999bc11fbbc049"} pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 05:08:21 crc kubenswrapper[4818]: I1122 05:08:21.266446 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" containerID="cri-o://7f4e621532f88c0da5dac82557bd8694f552742e853952ce47999bc11fbbc049" gracePeriod=600 Nov 22 05:08:21 crc kubenswrapper[4818]: I1122 05:08:21.615596 4818 generic.go:334] "Generic (PLEG): container finished" podID="0971c9ec-29ce-44eb-9875-ad4664143305" containerID="7f4e621532f88c0da5dac82557bd8694f552742e853952ce47999bc11fbbc049" exitCode=0 Nov 22 05:08:21 crc kubenswrapper[4818]: I1122 05:08:21.615660 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerDied","Data":"7f4e621532f88c0da5dac82557bd8694f552742e853952ce47999bc11fbbc049"} Nov 22 05:08:21 crc kubenswrapper[4818]: I1122 05:08:21.616025 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"f815386123126fabb0fbbe54df8d571044020c0a205a877246cd025618c83140"} Nov 22 05:08:21 crc kubenswrapper[4818]: I1122 05:08:21.616051 4818 scope.go:117] "RemoveContainer" containerID="17a029baa87e074670137f3b22e9f934cbfb0d28077e00586b8813874c289d90" Nov 22 05:08:21 crc kubenswrapper[4818]: I1122 05:08:21.618432 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-lx77w" event={"ID":"531548a2-b8d3-4a2a-a6e9-1bb2af1891d8","Type":"ContainerStarted","Data":"fabc8785f416eebcc5f4fd956d5a65463fe909641ad1fbba792da87cd0c5238e"} Nov 22 05:08:21 crc kubenswrapper[4818]: I1122 05:08:21.618644 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-lx77w" Nov 22 05:08:21 crc kubenswrapper[4818]: I1122 05:08:21.651392 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-lx77w" podStartSLOduration=4.160668949 podStartE2EDuration="1m8.651375782s" podCreationTimestamp="2025-11-22 05:07:13 +0000 UTC" firstStartedPulling="2025-11-22 05:07:16.224508477 +0000 UTC m=+1188.798925004" lastFinishedPulling="2025-11-22 05:08:20.71521531 +0000 UTC m=+1253.289631837" observedRunningTime="2025-11-22 05:08:21.648002891 +0000 UTC m=+1254.222419418" watchObservedRunningTime="2025-11-22 05:08:21.651375782 +0000 UTC m=+1254.225792309" Nov 22 05:08:24 crc kubenswrapper[4818]: I1122 05:08:24.650597 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ncfxv" event={"ID":"c31258b3-6cac-41e8-a350-1ea7b360b848","Type":"ContainerStarted","Data":"c06a61b9a163565cfb0981ede8b83a0b832b0ec18071a36ee2425afb703cd532"} Nov 22 05:08:24 crc kubenswrapper[4818]: I1122 05:08:24.651590 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ncfxv" Nov 22 05:08:24 crc kubenswrapper[4818]: I1122 05:08:24.679746 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ncfxv" podStartSLOduration=4.070874623 podStartE2EDuration="1m11.679695924s" podCreationTimestamp="2025-11-22 05:07:13 +0000 UTC" firstStartedPulling="2025-11-22 05:07:16.144554265 +0000 UTC m=+1188.718970792" lastFinishedPulling="2025-11-22 05:08:23.753375526 +0000 UTC m=+1256.327792093" observedRunningTime="2025-11-22 05:08:24.673195451 +0000 UTC m=+1257.247612028" watchObservedRunningTime="2025-11-22 05:08:24.679695924 +0000 UTC m=+1257.254112471" Nov 22 05:08:25 crc kubenswrapper[4818]: I1122 05:08:25.659355 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-56phb" event={"ID":"786ab73e-012b-47f6-b033-9fa9aab6d6cb","Type":"ContainerStarted","Data":"26e3e24e0296d208230de46a3fb4b71ecd1c7f7e13637baffa753d446f8073ad"} Nov 22 05:08:25 crc kubenswrapper[4818]: I1122 05:08:25.661236 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-56phb" Nov 22 05:08:25 crc kubenswrapper[4818]: I1122 05:08:25.663186 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d6957" event={"ID":"62c4e1cf-72fb-4a9b-88a1-350c1e20f360","Type":"ContainerStarted","Data":"d2a4788db146cb1dc55841f042ac08933968e955c1db197f7802850ffb73aa9d"} Nov 22 05:08:25 crc kubenswrapper[4818]: I1122 05:08:25.687893 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-56phb" podStartSLOduration=3.127556165 podStartE2EDuration="1m11.687860344s" podCreationTimestamp="2025-11-22 05:07:14 +0000 UTC" firstStartedPulling="2025-11-22 05:07:16.170848047 +0000 UTC m=+1188.745264574" lastFinishedPulling="2025-11-22 05:08:24.731152196 +0000 UTC m=+1257.305568753" observedRunningTime="2025-11-22 05:08:25.675950177 +0000 UTC m=+1258.250366714" watchObservedRunningTime="2025-11-22 05:08:25.687860344 +0000 UTC m=+1258.262276881" Nov 22 05:08:25 crc kubenswrapper[4818]: I1122 05:08:25.699003 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d6957" podStartSLOduration=2.990208933 podStartE2EDuration="1m11.698977871s" podCreationTimestamp="2025-11-22 05:07:14 +0000 UTC" firstStartedPulling="2025-11-22 05:07:16.105874974 +0000 UTC m=+1188.680291501" lastFinishedPulling="2025-11-22 05:08:24.814643892 +0000 UTC m=+1257.389060439" observedRunningTime="2025-11-22 05:08:25.68994239 +0000 UTC m=+1258.264358957" watchObservedRunningTime="2025-11-22 05:08:25.698977871 +0000 UTC m=+1258.273394408" Nov 22 05:08:27 crc kubenswrapper[4818]: I1122 05:08:27.684099 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-vkj22" event={"ID":"f04a965b-f5f4-45af-843f-862498956853","Type":"ContainerStarted","Data":"063bb58f08d464838121885243aed1b851dc7e73bb851f70ac3d04a900f30953"} Nov 22 05:08:27 crc kubenswrapper[4818]: I1122 05:08:27.684925 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d656998f4-vkj22" Nov 22 05:08:27 crc kubenswrapper[4818]: I1122 05:08:27.685599 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-pb45n" event={"ID":"e342ff02-d6d9-4be0-b97f-87dea6653a61","Type":"ContainerStarted","Data":"3033ae2e57aa1212956d8b5335dbffa780a4d9c36f887ae565fbdd2c88999761"} Nov 22 05:08:27 crc kubenswrapper[4818]: I1122 05:08:27.686224 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-b4c496f69-pb45n" Nov 22 05:08:27 crc kubenswrapper[4818]: I1122 05:08:27.704517 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-d656998f4-vkj22" podStartSLOduration=2.620052164 podStartE2EDuration="1m13.704494474s" podCreationTimestamp="2025-11-22 05:07:14 +0000 UTC" firstStartedPulling="2025-11-22 05:07:16.182390734 +0000 UTC m=+1188.756807261" lastFinishedPulling="2025-11-22 05:08:27.266833034 +0000 UTC m=+1259.841249571" observedRunningTime="2025-11-22 05:08:27.702912232 +0000 UTC m=+1260.277328759" watchObservedRunningTime="2025-11-22 05:08:27.704494474 +0000 UTC m=+1260.278911011" Nov 22 05:08:27 crc kubenswrapper[4818]: I1122 05:08:27.724825 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-b4c496f69-pb45n" podStartSLOduration=2.691976231 podStartE2EDuration="1m13.724803935s" podCreationTimestamp="2025-11-22 05:07:14 +0000 UTC" firstStartedPulling="2025-11-22 05:07:16.171142574 +0000 UTC m=+1188.745559101" lastFinishedPulling="2025-11-22 05:08:27.203970278 +0000 UTC m=+1259.778386805" observedRunningTime="2025-11-22 05:08:27.724346703 +0000 UTC m=+1260.298763240" watchObservedRunningTime="2025-11-22 05:08:27.724803935 +0000 UTC m=+1260.299220462" Nov 22 05:08:34 crc kubenswrapper[4818]: I1122 05:08:34.680132 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ncfxv" Nov 22 05:08:34 crc kubenswrapper[4818]: I1122 05:08:34.691832 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-56phb" Nov 22 05:08:34 crc kubenswrapper[4818]: I1122 05:08:34.723349 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d656998f4-vkj22" Nov 22 05:08:34 crc kubenswrapper[4818]: I1122 05:08:34.852699 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-lx77w" Nov 22 05:08:35 crc kubenswrapper[4818]: I1122 05:08:35.080994 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-b4c496f69-pb45n" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.110641 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nztvx"] Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.112878 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-nztvx" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.118224 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.118284 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.118377 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.118737 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-9hr45" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.119213 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nztvx"] Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.177470 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8p48\" (UniqueName: \"kubernetes.io/projected/ee2898b0-81b0-473b-959d-e19bcc3836cf-kube-api-access-s8p48\") pod \"dnsmasq-dns-675f4bcbfc-nztvx\" (UID: \"ee2898b0-81b0-473b-959d-e19bcc3836cf\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nztvx" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.177527 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee2898b0-81b0-473b-959d-e19bcc3836cf-config\") pod \"dnsmasq-dns-675f4bcbfc-nztvx\" (UID: \"ee2898b0-81b0-473b-959d-e19bcc3836cf\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nztvx" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.210028 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-t62q2"] Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.211223 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-t62q2" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.219047 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.241896 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-t62q2"] Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.279168 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f67a1621-8be8-490f-a2a5-489e11ba4b9e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-t62q2\" (UID: \"f67a1621-8be8-490f-a2a5-489e11ba4b9e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t62q2" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.279215 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk5ls\" (UniqueName: \"kubernetes.io/projected/f67a1621-8be8-490f-a2a5-489e11ba4b9e-kube-api-access-lk5ls\") pod \"dnsmasq-dns-78dd6ddcc-t62q2\" (UID: \"f67a1621-8be8-490f-a2a5-489e11ba4b9e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t62q2" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.279385 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8p48\" (UniqueName: \"kubernetes.io/projected/ee2898b0-81b0-473b-959d-e19bcc3836cf-kube-api-access-s8p48\") pod \"dnsmasq-dns-675f4bcbfc-nztvx\" (UID: \"ee2898b0-81b0-473b-959d-e19bcc3836cf\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nztvx" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.279448 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee2898b0-81b0-473b-959d-e19bcc3836cf-config\") pod \"dnsmasq-dns-675f4bcbfc-nztvx\" (UID: \"ee2898b0-81b0-473b-959d-e19bcc3836cf\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nztvx" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.279497 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f67a1621-8be8-490f-a2a5-489e11ba4b9e-config\") pod \"dnsmasq-dns-78dd6ddcc-t62q2\" (UID: \"f67a1621-8be8-490f-a2a5-489e11ba4b9e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t62q2" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.280731 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee2898b0-81b0-473b-959d-e19bcc3836cf-config\") pod \"dnsmasq-dns-675f4bcbfc-nztvx\" (UID: \"ee2898b0-81b0-473b-959d-e19bcc3836cf\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nztvx" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.296952 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8p48\" (UniqueName: \"kubernetes.io/projected/ee2898b0-81b0-473b-959d-e19bcc3836cf-kube-api-access-s8p48\") pod \"dnsmasq-dns-675f4bcbfc-nztvx\" (UID: \"ee2898b0-81b0-473b-959d-e19bcc3836cf\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nztvx" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.380505 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f67a1621-8be8-490f-a2a5-489e11ba4b9e-config\") pod \"dnsmasq-dns-78dd6ddcc-t62q2\" (UID: \"f67a1621-8be8-490f-a2a5-489e11ba4b9e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t62q2" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.380588 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f67a1621-8be8-490f-a2a5-489e11ba4b9e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-t62q2\" (UID: \"f67a1621-8be8-490f-a2a5-489e11ba4b9e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t62q2" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.380608 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk5ls\" (UniqueName: \"kubernetes.io/projected/f67a1621-8be8-490f-a2a5-489e11ba4b9e-kube-api-access-lk5ls\") pod \"dnsmasq-dns-78dd6ddcc-t62q2\" (UID: \"f67a1621-8be8-490f-a2a5-489e11ba4b9e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t62q2" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.381524 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f67a1621-8be8-490f-a2a5-489e11ba4b9e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-t62q2\" (UID: \"f67a1621-8be8-490f-a2a5-489e11ba4b9e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t62q2" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.381562 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f67a1621-8be8-490f-a2a5-489e11ba4b9e-config\") pod \"dnsmasq-dns-78dd6ddcc-t62q2\" (UID: \"f67a1621-8be8-490f-a2a5-489e11ba4b9e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t62q2" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.395516 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk5ls\" (UniqueName: \"kubernetes.io/projected/f67a1621-8be8-490f-a2a5-489e11ba4b9e-kube-api-access-lk5ls\") pod \"dnsmasq-dns-78dd6ddcc-t62q2\" (UID: \"f67a1621-8be8-490f-a2a5-489e11ba4b9e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t62q2" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.471795 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-nztvx" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.527789 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-t62q2" Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.909393 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nztvx"] Nov 22 05:08:52 crc kubenswrapper[4818]: I1122 05:08:52.981551 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-t62q2"] Nov 22 05:08:52 crc kubenswrapper[4818]: W1122 05:08:52.984175 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf67a1621_8be8_490f_a2a5_489e11ba4b9e.slice/crio-89dcf2ffdb36837345685292f59d27013b3402e6166212c021a8952bcc32fc85 WatchSource:0}: Error finding container 89dcf2ffdb36837345685292f59d27013b3402e6166212c021a8952bcc32fc85: Status 404 returned error can't find the container with id 89dcf2ffdb36837345685292f59d27013b3402e6166212c021a8952bcc32fc85 Nov 22 05:08:53 crc kubenswrapper[4818]: I1122 05:08:53.923275 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-nztvx" event={"ID":"ee2898b0-81b0-473b-959d-e19bcc3836cf","Type":"ContainerStarted","Data":"8a653a934cc385cfd6eb4e907c0c02baa9bc5a73569e469d943df9429fde3a02"} Nov 22 05:08:53 crc kubenswrapper[4818]: I1122 05:08:53.925307 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-t62q2" event={"ID":"f67a1621-8be8-490f-a2a5-489e11ba4b9e","Type":"ContainerStarted","Data":"89dcf2ffdb36837345685292f59d27013b3402e6166212c021a8952bcc32fc85"} Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.175705 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nztvx"] Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.203536 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-j8tk7"] Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.204777 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.209900 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-j8tk7"] Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.339032 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/acd97d29-b853-4de1-87e6-54d68a856eeb-dns-svc\") pod \"dnsmasq-dns-666b6646f7-j8tk7\" (UID: \"acd97d29-b853-4de1-87e6-54d68a856eeb\") " pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.339086 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdf7f\" (UniqueName: \"kubernetes.io/projected/acd97d29-b853-4de1-87e6-54d68a856eeb-kube-api-access-sdf7f\") pod \"dnsmasq-dns-666b6646f7-j8tk7\" (UID: \"acd97d29-b853-4de1-87e6-54d68a856eeb\") " pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.339116 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acd97d29-b853-4de1-87e6-54d68a856eeb-config\") pod \"dnsmasq-dns-666b6646f7-j8tk7\" (UID: \"acd97d29-b853-4de1-87e6-54d68a856eeb\") " pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.440784 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/acd97d29-b853-4de1-87e6-54d68a856eeb-dns-svc\") pod \"dnsmasq-dns-666b6646f7-j8tk7\" (UID: \"acd97d29-b853-4de1-87e6-54d68a856eeb\") " pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.441086 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdf7f\" (UniqueName: \"kubernetes.io/projected/acd97d29-b853-4de1-87e6-54d68a856eeb-kube-api-access-sdf7f\") pod \"dnsmasq-dns-666b6646f7-j8tk7\" (UID: \"acd97d29-b853-4de1-87e6-54d68a856eeb\") " pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.441113 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acd97d29-b853-4de1-87e6-54d68a856eeb-config\") pod \"dnsmasq-dns-666b6646f7-j8tk7\" (UID: \"acd97d29-b853-4de1-87e6-54d68a856eeb\") " pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.442095 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acd97d29-b853-4de1-87e6-54d68a856eeb-config\") pod \"dnsmasq-dns-666b6646f7-j8tk7\" (UID: \"acd97d29-b853-4de1-87e6-54d68a856eeb\") " pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.442687 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/acd97d29-b853-4de1-87e6-54d68a856eeb-dns-svc\") pod \"dnsmasq-dns-666b6646f7-j8tk7\" (UID: \"acd97d29-b853-4de1-87e6-54d68a856eeb\") " pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.494595 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdf7f\" (UniqueName: \"kubernetes.io/projected/acd97d29-b853-4de1-87e6-54d68a856eeb-kube-api-access-sdf7f\") pod \"dnsmasq-dns-666b6646f7-j8tk7\" (UID: \"acd97d29-b853-4de1-87e6-54d68a856eeb\") " pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.523630 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.544504 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-t62q2"] Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.591063 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-7c9sg"] Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.592310 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.626416 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-7c9sg"] Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.644458 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvcnk\" (UniqueName: \"kubernetes.io/projected/9bfecc94-1073-46f1-9ea1-378ddc74ecf6-kube-api-access-rvcnk\") pod \"dnsmasq-dns-57d769cc4f-7c9sg\" (UID: \"9bfecc94-1073-46f1-9ea1-378ddc74ecf6\") " pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.644939 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bfecc94-1073-46f1-9ea1-378ddc74ecf6-config\") pod \"dnsmasq-dns-57d769cc4f-7c9sg\" (UID: \"9bfecc94-1073-46f1-9ea1-378ddc74ecf6\") " pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.645101 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bfecc94-1073-46f1-9ea1-378ddc74ecf6-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-7c9sg\" (UID: \"9bfecc94-1073-46f1-9ea1-378ddc74ecf6\") " pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.746772 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bfecc94-1073-46f1-9ea1-378ddc74ecf6-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-7c9sg\" (UID: \"9bfecc94-1073-46f1-9ea1-378ddc74ecf6\") " pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.747133 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvcnk\" (UniqueName: \"kubernetes.io/projected/9bfecc94-1073-46f1-9ea1-378ddc74ecf6-kube-api-access-rvcnk\") pod \"dnsmasq-dns-57d769cc4f-7c9sg\" (UID: \"9bfecc94-1073-46f1-9ea1-378ddc74ecf6\") " pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.747214 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bfecc94-1073-46f1-9ea1-378ddc74ecf6-config\") pod \"dnsmasq-dns-57d769cc4f-7c9sg\" (UID: \"9bfecc94-1073-46f1-9ea1-378ddc74ecf6\") " pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.748952 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bfecc94-1073-46f1-9ea1-378ddc74ecf6-config\") pod \"dnsmasq-dns-57d769cc4f-7c9sg\" (UID: \"9bfecc94-1073-46f1-9ea1-378ddc74ecf6\") " pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.765691 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bfecc94-1073-46f1-9ea1-378ddc74ecf6-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-7c9sg\" (UID: \"9bfecc94-1073-46f1-9ea1-378ddc74ecf6\") " pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.772953 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvcnk\" (UniqueName: \"kubernetes.io/projected/9bfecc94-1073-46f1-9ea1-378ddc74ecf6-kube-api-access-rvcnk\") pod \"dnsmasq-dns-57d769cc4f-7c9sg\" (UID: \"9bfecc94-1073-46f1-9ea1-378ddc74ecf6\") " pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" Nov 22 05:08:55 crc kubenswrapper[4818]: I1122 05:08:55.957292 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.090528 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-j8tk7"] Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.332513 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.333940 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.336140 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.336221 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.336449 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.336612 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-26pzg" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.336736 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.337041 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.337197 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.352296 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.457923 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/69f72526-235a-4079-a089-43e1b93aef9f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.457964 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.458008 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.458100 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/69f72526-235a-4079-a089-43e1b93aef9f-config-data\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.458172 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/69f72526-235a-4079-a089-43e1b93aef9f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.458201 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/69f72526-235a-4079-a089-43e1b93aef9f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.458492 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.459039 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/69f72526-235a-4079-a089-43e1b93aef9f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.459298 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.459528 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9dn6\" (UniqueName: \"kubernetes.io/projected/69f72526-235a-4079-a089-43e1b93aef9f-kube-api-access-f9dn6\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.459618 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.560921 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.560964 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/69f72526-235a-4079-a089-43e1b93aef9f-config-data\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.560992 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/69f72526-235a-4079-a089-43e1b93aef9f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.561013 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/69f72526-235a-4079-a089-43e1b93aef9f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.561035 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.561071 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/69f72526-235a-4079-a089-43e1b93aef9f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.561106 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.561136 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9dn6\" (UniqueName: \"kubernetes.io/projected/69f72526-235a-4079-a089-43e1b93aef9f-kube-api-access-f9dn6\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.561154 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.561175 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/69f72526-235a-4079-a089-43e1b93aef9f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.561192 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.563020 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.563589 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.563785 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/69f72526-235a-4079-a089-43e1b93aef9f-config-data\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.563986 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.564485 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/69f72526-235a-4079-a089-43e1b93aef9f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.565966 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/69f72526-235a-4079-a089-43e1b93aef9f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.567784 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.567784 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/69f72526-235a-4079-a089-43e1b93aef9f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.568318 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.577721 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/69f72526-235a-4079-a089-43e1b93aef9f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.579919 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9dn6\" (UniqueName: \"kubernetes.io/projected/69f72526-235a-4079-a089-43e1b93aef9f-kube-api-access-f9dn6\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.586760 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.661606 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.699520 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.700688 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.703876 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.703960 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.704045 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.704115 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.704148 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.703961 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-cws5h" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.707235 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.712743 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.864547 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/67e09220-fd61-4785-9de7-90b3ffaf0157-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.864591 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.864624 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.864655 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqs9r\" (UniqueName: \"kubernetes.io/projected/67e09220-fd61-4785-9de7-90b3ffaf0157-kube-api-access-bqs9r\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.864774 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/67e09220-fd61-4785-9de7-90b3ffaf0157-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.864873 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.865027 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/67e09220-fd61-4785-9de7-90b3ffaf0157-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.865146 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/67e09220-fd61-4785-9de7-90b3ffaf0157-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.865200 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/67e09220-fd61-4785-9de7-90b3ffaf0157-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.865347 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.865385 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.966587 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.967009 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/67e09220-fd61-4785-9de7-90b3ffaf0157-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.967054 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/67e09220-fd61-4785-9de7-90b3ffaf0157-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.967080 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/67e09220-fd61-4785-9de7-90b3ffaf0157-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.967139 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.967165 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.967197 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/67e09220-fd61-4785-9de7-90b3ffaf0157-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.967228 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.967273 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.967298 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqs9r\" (UniqueName: \"kubernetes.io/projected/67e09220-fd61-4785-9de7-90b3ffaf0157-kube-api-access-bqs9r\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.967322 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/67e09220-fd61-4785-9de7-90b3ffaf0157-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.967655 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.968230 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/67e09220-fd61-4785-9de7-90b3ffaf0157-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.968865 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/67e09220-fd61-4785-9de7-90b3ffaf0157-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.971168 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.971592 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/67e09220-fd61-4785-9de7-90b3ffaf0157-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.973738 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.974300 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.975005 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/67e09220-fd61-4785-9de7-90b3ffaf0157-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.975061 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/67e09220-fd61-4785-9de7-90b3ffaf0157-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.975718 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:56 crc kubenswrapper[4818]: I1122 05:08:56.997776 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqs9r\" (UniqueName: \"kubernetes.io/projected/67e09220-fd61-4785-9de7-90b3ffaf0157-kube-api-access-bqs9r\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:57 crc kubenswrapper[4818]: I1122 05:08:57.008684 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:57 crc kubenswrapper[4818]: I1122 05:08:57.027027 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.032613 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.036633 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.042194 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.042322 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.042531 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-7fvnt" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.043182 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.047294 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.058296 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.197317 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4714282b-39ba-4691-8fbd-c3e76120fbb5-kolla-config\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.197359 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nrm4\" (UniqueName: \"kubernetes.io/projected/4714282b-39ba-4691-8fbd-c3e76120fbb5-kube-api-access-9nrm4\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.197394 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.197418 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4714282b-39ba-4691-8fbd-c3e76120fbb5-config-data-generated\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.197477 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4714282b-39ba-4691-8fbd-c3e76120fbb5-operator-scripts\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.197509 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4714282b-39ba-4691-8fbd-c3e76120fbb5-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.197531 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4714282b-39ba-4691-8fbd-c3e76120fbb5-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.197572 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4714282b-39ba-4691-8fbd-c3e76120fbb5-config-data-default\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.299423 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4714282b-39ba-4691-8fbd-c3e76120fbb5-operator-scripts\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.299461 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4714282b-39ba-4691-8fbd-c3e76120fbb5-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.299489 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4714282b-39ba-4691-8fbd-c3e76120fbb5-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.299526 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4714282b-39ba-4691-8fbd-c3e76120fbb5-config-data-default\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.299581 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4714282b-39ba-4691-8fbd-c3e76120fbb5-kolla-config\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.299598 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nrm4\" (UniqueName: \"kubernetes.io/projected/4714282b-39ba-4691-8fbd-c3e76120fbb5-kube-api-access-9nrm4\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.299626 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.299657 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4714282b-39ba-4691-8fbd-c3e76120fbb5-config-data-generated\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.300068 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4714282b-39ba-4691-8fbd-c3e76120fbb5-config-data-generated\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.300448 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4714282b-39ba-4691-8fbd-c3e76120fbb5-config-data-default\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.300611 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.300754 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4714282b-39ba-4691-8fbd-c3e76120fbb5-kolla-config\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.301125 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4714282b-39ba-4691-8fbd-c3e76120fbb5-operator-scripts\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.304726 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4714282b-39ba-4691-8fbd-c3e76120fbb5-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.315004 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4714282b-39ba-4691-8fbd-c3e76120fbb5-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.321637 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nrm4\" (UniqueName: \"kubernetes.io/projected/4714282b-39ba-4691-8fbd-c3e76120fbb5-kube-api-access-9nrm4\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.339623 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"4714282b-39ba-4691-8fbd-c3e76120fbb5\") " pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.368539 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 22 05:08:58 crc kubenswrapper[4818]: I1122 05:08:58.990654 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" event={"ID":"acd97d29-b853-4de1-87e6-54d68a856eeb","Type":"ContainerStarted","Data":"7499d439594b45be495f421f475caa694a5e85b38a1f62a7416b1eeed941110c"} Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.599777 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.600981 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.604060 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.604101 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.604369 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-npgts" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.618464 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.631879 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.723545 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/72cac26c-ba82-41c7-a405-a9162867abea-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.723590 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/72cac26c-ba82-41c7-a405-a9162867abea-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.723612 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npwgp\" (UniqueName: \"kubernetes.io/projected/72cac26c-ba82-41c7-a405-a9162867abea-kube-api-access-npwgp\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.723638 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.723657 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72cac26c-ba82-41c7-a405-a9162867abea-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.723677 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/72cac26c-ba82-41c7-a405-a9162867abea-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.723842 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72cac26c-ba82-41c7-a405-a9162867abea-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.723859 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/72cac26c-ba82-41c7-a405-a9162867abea-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.825365 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/72cac26c-ba82-41c7-a405-a9162867abea-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.825424 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/72cac26c-ba82-41c7-a405-a9162867abea-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.825452 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npwgp\" (UniqueName: \"kubernetes.io/projected/72cac26c-ba82-41c7-a405-a9162867abea-kube-api-access-npwgp\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.825480 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.825501 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72cac26c-ba82-41c7-a405-a9162867abea-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.825522 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/72cac26c-ba82-41c7-a405-a9162867abea-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.825560 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/72cac26c-ba82-41c7-a405-a9162867abea-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.825577 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72cac26c-ba82-41c7-a405-a9162867abea-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.827638 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/72cac26c-ba82-41c7-a405-a9162867abea-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.828040 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/72cac26c-ba82-41c7-a405-a9162867abea-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.828113 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/72cac26c-ba82-41c7-a405-a9162867abea-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.828203 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.828465 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72cac26c-ba82-41c7-a405-a9162867abea-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.842296 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72cac26c-ba82-41c7-a405-a9162867abea-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.844655 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/72cac26c-ba82-41c7-a405-a9162867abea-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.849485 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.856412 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npwgp\" (UniqueName: \"kubernetes.io/projected/72cac26c-ba82-41c7-a405-a9162867abea-kube-api-access-npwgp\") pod \"openstack-cell1-galera-0\" (UID: \"72cac26c-ba82-41c7-a405-a9162867abea\") " pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.951539 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.996851 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 22 05:08:59 crc kubenswrapper[4818]: I1122 05:08:59.998026 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 22 05:09:00 crc kubenswrapper[4818]: I1122 05:09:00.001634 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 22 05:09:00 crc kubenswrapper[4818]: I1122 05:09:00.001855 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-2vjb7" Nov 22 05:09:00 crc kubenswrapper[4818]: I1122 05:09:00.001897 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 22 05:09:00 crc kubenswrapper[4818]: I1122 05:09:00.013968 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 22 05:09:00 crc kubenswrapper[4818]: I1122 05:09:00.130513 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e98d9180-5e2b-4b38-8fc0-844072603c2d-config-data\") pod \"memcached-0\" (UID: \"e98d9180-5e2b-4b38-8fc0-844072603c2d\") " pod="openstack/memcached-0" Nov 22 05:09:00 crc kubenswrapper[4818]: I1122 05:09:00.130565 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e98d9180-5e2b-4b38-8fc0-844072603c2d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e98d9180-5e2b-4b38-8fc0-844072603c2d\") " pod="openstack/memcached-0" Nov 22 05:09:00 crc kubenswrapper[4818]: I1122 05:09:00.130598 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e98d9180-5e2b-4b38-8fc0-844072603c2d-kolla-config\") pod \"memcached-0\" (UID: \"e98d9180-5e2b-4b38-8fc0-844072603c2d\") " pod="openstack/memcached-0" Nov 22 05:09:00 crc kubenswrapper[4818]: I1122 05:09:00.130742 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e98d9180-5e2b-4b38-8fc0-844072603c2d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e98d9180-5e2b-4b38-8fc0-844072603c2d\") " pod="openstack/memcached-0" Nov 22 05:09:00 crc kubenswrapper[4818]: I1122 05:09:00.130933 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk9xx\" (UniqueName: \"kubernetes.io/projected/e98d9180-5e2b-4b38-8fc0-844072603c2d-kube-api-access-hk9xx\") pod \"memcached-0\" (UID: \"e98d9180-5e2b-4b38-8fc0-844072603c2d\") " pod="openstack/memcached-0" Nov 22 05:09:00 crc kubenswrapper[4818]: I1122 05:09:00.232494 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk9xx\" (UniqueName: \"kubernetes.io/projected/e98d9180-5e2b-4b38-8fc0-844072603c2d-kube-api-access-hk9xx\") pod \"memcached-0\" (UID: \"e98d9180-5e2b-4b38-8fc0-844072603c2d\") " pod="openstack/memcached-0" Nov 22 05:09:00 crc kubenswrapper[4818]: I1122 05:09:00.232604 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e98d9180-5e2b-4b38-8fc0-844072603c2d-config-data\") pod \"memcached-0\" (UID: \"e98d9180-5e2b-4b38-8fc0-844072603c2d\") " pod="openstack/memcached-0" Nov 22 05:09:00 crc kubenswrapper[4818]: I1122 05:09:00.232639 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e98d9180-5e2b-4b38-8fc0-844072603c2d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e98d9180-5e2b-4b38-8fc0-844072603c2d\") " pod="openstack/memcached-0" Nov 22 05:09:00 crc kubenswrapper[4818]: I1122 05:09:00.232676 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e98d9180-5e2b-4b38-8fc0-844072603c2d-kolla-config\") pod \"memcached-0\" (UID: \"e98d9180-5e2b-4b38-8fc0-844072603c2d\") " pod="openstack/memcached-0" Nov 22 05:09:00 crc kubenswrapper[4818]: I1122 05:09:00.232708 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e98d9180-5e2b-4b38-8fc0-844072603c2d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e98d9180-5e2b-4b38-8fc0-844072603c2d\") " pod="openstack/memcached-0" Nov 22 05:09:00 crc kubenswrapper[4818]: I1122 05:09:00.233923 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e98d9180-5e2b-4b38-8fc0-844072603c2d-config-data\") pod \"memcached-0\" (UID: \"e98d9180-5e2b-4b38-8fc0-844072603c2d\") " pod="openstack/memcached-0" Nov 22 05:09:00 crc kubenswrapper[4818]: I1122 05:09:00.234714 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e98d9180-5e2b-4b38-8fc0-844072603c2d-kolla-config\") pod \"memcached-0\" (UID: \"e98d9180-5e2b-4b38-8fc0-844072603c2d\") " pod="openstack/memcached-0" Nov 22 05:09:00 crc kubenswrapper[4818]: I1122 05:09:00.236345 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e98d9180-5e2b-4b38-8fc0-844072603c2d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e98d9180-5e2b-4b38-8fc0-844072603c2d\") " pod="openstack/memcached-0" Nov 22 05:09:00 crc kubenswrapper[4818]: I1122 05:09:00.244100 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e98d9180-5e2b-4b38-8fc0-844072603c2d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e98d9180-5e2b-4b38-8fc0-844072603c2d\") " pod="openstack/memcached-0" Nov 22 05:09:00 crc kubenswrapper[4818]: I1122 05:09:00.261288 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk9xx\" (UniqueName: \"kubernetes.io/projected/e98d9180-5e2b-4b38-8fc0-844072603c2d-kube-api-access-hk9xx\") pod \"memcached-0\" (UID: \"e98d9180-5e2b-4b38-8fc0-844072603c2d\") " pod="openstack/memcached-0" Nov 22 05:09:00 crc kubenswrapper[4818]: I1122 05:09:00.317914 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 22 05:09:01 crc kubenswrapper[4818]: I1122 05:09:01.622920 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 05:09:01 crc kubenswrapper[4818]: I1122 05:09:01.624129 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 22 05:09:01 crc kubenswrapper[4818]: I1122 05:09:01.625973 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-6qhxv" Nov 22 05:09:01 crc kubenswrapper[4818]: I1122 05:09:01.632648 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 05:09:01 crc kubenswrapper[4818]: I1122 05:09:01.752571 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld5wh\" (UniqueName: \"kubernetes.io/projected/92a5743a-8fbe-49a7-a653-2b32e6034f52-kube-api-access-ld5wh\") pod \"kube-state-metrics-0\" (UID: \"92a5743a-8fbe-49a7-a653-2b32e6034f52\") " pod="openstack/kube-state-metrics-0" Nov 22 05:09:01 crc kubenswrapper[4818]: I1122 05:09:01.854321 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld5wh\" (UniqueName: \"kubernetes.io/projected/92a5743a-8fbe-49a7-a653-2b32e6034f52-kube-api-access-ld5wh\") pod \"kube-state-metrics-0\" (UID: \"92a5743a-8fbe-49a7-a653-2b32e6034f52\") " pod="openstack/kube-state-metrics-0" Nov 22 05:09:01 crc kubenswrapper[4818]: I1122 05:09:01.874870 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld5wh\" (UniqueName: \"kubernetes.io/projected/92a5743a-8fbe-49a7-a653-2b32e6034f52-kube-api-access-ld5wh\") pod \"kube-state-metrics-0\" (UID: \"92a5743a-8fbe-49a7-a653-2b32e6034f52\") " pod="openstack/kube-state-metrics-0" Nov 22 05:09:01 crc kubenswrapper[4818]: I1122 05:09:01.947709 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.176398 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jgq5d"] Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.177953 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.180038 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.183725 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-kpdh5"] Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.184081 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-jhtjk" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.184107 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.188763 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.189614 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jgq5d"] Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.207853 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-kpdh5"] Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.335480 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1b31a0d3-3489-430a-853a-40d8715f7e7f-var-log\") pod \"ovn-controller-ovs-kpdh5\" (UID: \"1b31a0d3-3489-430a-853a-40d8715f7e7f\") " pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.335529 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1b31a0d3-3489-430a-853a-40d8715f7e7f-var-run\") pod \"ovn-controller-ovs-kpdh5\" (UID: \"1b31a0d3-3489-430a-853a-40d8715f7e7f\") " pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.335551 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b31a0d3-3489-430a-853a-40d8715f7e7f-scripts\") pod \"ovn-controller-ovs-kpdh5\" (UID: \"1b31a0d3-3489-430a-853a-40d8715f7e7f\") " pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.335578 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7dec9ecf-e438-4334-a20e-757ce6fff445-var-run-ovn\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.335797 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7dec9ecf-e438-4334-a20e-757ce6fff445-var-run\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.335845 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7dec9ecf-e438-4334-a20e-757ce6fff445-scripts\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.335976 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1b31a0d3-3489-430a-853a-40d8715f7e7f-etc-ovs\") pod \"ovn-controller-ovs-kpdh5\" (UID: \"1b31a0d3-3489-430a-853a-40d8715f7e7f\") " pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.336021 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dec9ecf-e438-4334-a20e-757ce6fff445-combined-ca-bundle\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.336053 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7dec9ecf-e438-4334-a20e-757ce6fff445-var-log-ovn\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.336077 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1b31a0d3-3489-430a-853a-40d8715f7e7f-var-lib\") pod \"ovn-controller-ovs-kpdh5\" (UID: \"1b31a0d3-3489-430a-853a-40d8715f7e7f\") " pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.336094 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrclb\" (UniqueName: \"kubernetes.io/projected/1b31a0d3-3489-430a-853a-40d8715f7e7f-kube-api-access-vrclb\") pod \"ovn-controller-ovs-kpdh5\" (UID: \"1b31a0d3-3489-430a-853a-40d8715f7e7f\") " pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.336205 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dec9ecf-e438-4334-a20e-757ce6fff445-ovn-controller-tls-certs\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.336275 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfj6d\" (UniqueName: \"kubernetes.io/projected/7dec9ecf-e438-4334-a20e-757ce6fff445-kube-api-access-zfj6d\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.437622 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7dec9ecf-e438-4334-a20e-757ce6fff445-var-log-ovn\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.437660 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1b31a0d3-3489-430a-853a-40d8715f7e7f-var-lib\") pod \"ovn-controller-ovs-kpdh5\" (UID: \"1b31a0d3-3489-430a-853a-40d8715f7e7f\") " pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.437692 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrclb\" (UniqueName: \"kubernetes.io/projected/1b31a0d3-3489-430a-853a-40d8715f7e7f-kube-api-access-vrclb\") pod \"ovn-controller-ovs-kpdh5\" (UID: \"1b31a0d3-3489-430a-853a-40d8715f7e7f\") " pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.437725 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dec9ecf-e438-4334-a20e-757ce6fff445-ovn-controller-tls-certs\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.437746 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfj6d\" (UniqueName: \"kubernetes.io/projected/7dec9ecf-e438-4334-a20e-757ce6fff445-kube-api-access-zfj6d\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.437792 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1b31a0d3-3489-430a-853a-40d8715f7e7f-var-log\") pod \"ovn-controller-ovs-kpdh5\" (UID: \"1b31a0d3-3489-430a-853a-40d8715f7e7f\") " pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.437807 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1b31a0d3-3489-430a-853a-40d8715f7e7f-var-run\") pod \"ovn-controller-ovs-kpdh5\" (UID: \"1b31a0d3-3489-430a-853a-40d8715f7e7f\") " pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.437822 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b31a0d3-3489-430a-853a-40d8715f7e7f-scripts\") pod \"ovn-controller-ovs-kpdh5\" (UID: \"1b31a0d3-3489-430a-853a-40d8715f7e7f\") " pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.437840 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7dec9ecf-e438-4334-a20e-757ce6fff445-var-run-ovn\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.437886 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7dec9ecf-e438-4334-a20e-757ce6fff445-var-run\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.437905 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7dec9ecf-e438-4334-a20e-757ce6fff445-scripts\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.437936 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1b31a0d3-3489-430a-853a-40d8715f7e7f-etc-ovs\") pod \"ovn-controller-ovs-kpdh5\" (UID: \"1b31a0d3-3489-430a-853a-40d8715f7e7f\") " pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.437956 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dec9ecf-e438-4334-a20e-757ce6fff445-combined-ca-bundle\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.438188 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7dec9ecf-e438-4334-a20e-757ce6fff445-var-log-ovn\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.439221 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1b31a0d3-3489-430a-853a-40d8715f7e7f-var-lib\") pod \"ovn-controller-ovs-kpdh5\" (UID: \"1b31a0d3-3489-430a-853a-40d8715f7e7f\") " pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.439414 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1b31a0d3-3489-430a-853a-40d8715f7e7f-etc-ovs\") pod \"ovn-controller-ovs-kpdh5\" (UID: \"1b31a0d3-3489-430a-853a-40d8715f7e7f\") " pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.439457 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7dec9ecf-e438-4334-a20e-757ce6fff445-var-run-ovn\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.439591 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1b31a0d3-3489-430a-853a-40d8715f7e7f-var-log\") pod \"ovn-controller-ovs-kpdh5\" (UID: \"1b31a0d3-3489-430a-853a-40d8715f7e7f\") " pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.439693 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7dec9ecf-e438-4334-a20e-757ce6fff445-var-run\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.439740 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1b31a0d3-3489-430a-853a-40d8715f7e7f-var-run\") pod \"ovn-controller-ovs-kpdh5\" (UID: \"1b31a0d3-3489-430a-853a-40d8715f7e7f\") " pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.442398 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7dec9ecf-e438-4334-a20e-757ce6fff445-scripts\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.442460 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b31a0d3-3489-430a-853a-40d8715f7e7f-scripts\") pod \"ovn-controller-ovs-kpdh5\" (UID: \"1b31a0d3-3489-430a-853a-40d8715f7e7f\") " pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.444372 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dec9ecf-e438-4334-a20e-757ce6fff445-combined-ca-bundle\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.444719 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dec9ecf-e438-4334-a20e-757ce6fff445-ovn-controller-tls-certs\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.463006 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfj6d\" (UniqueName: \"kubernetes.io/projected/7dec9ecf-e438-4334-a20e-757ce6fff445-kube-api-access-zfj6d\") pod \"ovn-controller-jgq5d\" (UID: \"7dec9ecf-e438-4334-a20e-757ce6fff445\") " pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.477499 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrclb\" (UniqueName: \"kubernetes.io/projected/1b31a0d3-3489-430a-853a-40d8715f7e7f-kube-api-access-vrclb\") pod \"ovn-controller-ovs-kpdh5\" (UID: \"1b31a0d3-3489-430a-853a-40d8715f7e7f\") " pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.511111 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:06 crc kubenswrapper[4818]: I1122 05:09:06.528671 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.431398 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.433012 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.435757 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.435773 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.436200 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.436584 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.436777 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-fcmbl" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.449826 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.586180 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt8n5\" (UniqueName: \"kubernetes.io/projected/52d9e522-f84d-4671-9981-d08b3d7e0bbf-kube-api-access-dt8n5\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.586522 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/52d9e522-f84d-4671-9981-d08b3d7e0bbf-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.586560 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/52d9e522-f84d-4671-9981-d08b3d7e0bbf-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.586580 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52d9e522-f84d-4671-9981-d08b3d7e0bbf-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.586603 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52d9e522-f84d-4671-9981-d08b3d7e0bbf-config\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.586739 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.586756 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/52d9e522-f84d-4671-9981-d08b3d7e0bbf-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.586782 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/52d9e522-f84d-4671-9981-d08b3d7e0bbf-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.647991 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.649207 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.655708 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-5n2vc" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.655934 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.658538 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.659966 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.664679 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.687835 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt8n5\" (UniqueName: \"kubernetes.io/projected/52d9e522-f84d-4671-9981-d08b3d7e0bbf-kube-api-access-dt8n5\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.687890 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/52d9e522-f84d-4671-9981-d08b3d7e0bbf-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.687929 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/52d9e522-f84d-4671-9981-d08b3d7e0bbf-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.687947 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52d9e522-f84d-4671-9981-d08b3d7e0bbf-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.687970 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52d9e522-f84d-4671-9981-d08b3d7e0bbf-config\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.688020 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.688034 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/52d9e522-f84d-4671-9981-d08b3d7e0bbf-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.688060 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/52d9e522-f84d-4671-9981-d08b3d7e0bbf-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.689190 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.689428 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/52d9e522-f84d-4671-9981-d08b3d7e0bbf-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.690176 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/52d9e522-f84d-4671-9981-d08b3d7e0bbf-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.690773 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52d9e522-f84d-4671-9981-d08b3d7e0bbf-config\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.694525 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52d9e522-f84d-4671-9981-d08b3d7e0bbf-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.697539 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/52d9e522-f84d-4671-9981-d08b3d7e0bbf-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.699966 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/52d9e522-f84d-4671-9981-d08b3d7e0bbf-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.718846 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt8n5\" (UniqueName: \"kubernetes.io/projected/52d9e522-f84d-4671-9981-d08b3d7e0bbf-kube-api-access-dt8n5\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.719865 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"52d9e522-f84d-4671-9981-d08b3d7e0bbf\") " pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.790207 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.790299 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8ns9\" (UniqueName: \"kubernetes.io/projected/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-kube-api-access-p8ns9\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.790340 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.790378 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.790409 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.790433 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.790505 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-config\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.790530 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.808877 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.891458 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-config\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.891497 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.891552 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.891594 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8ns9\" (UniqueName: \"kubernetes.io/projected/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-kube-api-access-p8ns9\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.891624 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.891658 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.891686 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.891711 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.892601 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.893030 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-config\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.893184 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.894467 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.896963 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.897155 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.898362 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.918585 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8ns9\" (UniqueName: \"kubernetes.io/projected/5b0746b7-8400-44c8-b4c5-c6b6e3811e91-kube-api-access-p8ns9\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.920924 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"5b0746b7-8400-44c8-b4c5-c6b6e3811e91\") " pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:08 crc kubenswrapper[4818]: I1122 05:09:08.961574 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:21 crc kubenswrapper[4818]: E1122 05:09:21.764540 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = writing blob: storing blob to file \"/var/tmp/container_images_storage2631169349/1\": happened during read: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 22 05:09:21 crc kubenswrapper[4818]: E1122 05:09:21.765207 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sdf7f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-j8tk7_openstack(acd97d29-b853-4de1-87e6-54d68a856eeb): ErrImagePull: rpc error: code = Canceled desc = writing blob: storing blob to file \"/var/tmp/container_images_storage2631169349/1\": happened during read: context canceled" logger="UnhandledError" Nov 22 05:09:21 crc kubenswrapper[4818]: E1122 05:09:21.766668 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = writing blob: storing blob to file \\\"/var/tmp/container_images_storage2631169349/1\\\": happened during read: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" podUID="acd97d29-b853-4de1-87e6-54d68a856eeb" Nov 22 05:09:22 crc kubenswrapper[4818]: E1122 05:09:22.192991 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" podUID="acd97d29-b853-4de1-87e6-54d68a856eeb" Nov 22 05:09:22 crc kubenswrapper[4818]: I1122 05:09:22.211958 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 05:09:22 crc kubenswrapper[4818]: W1122 05:09:22.213417 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67e09220_fd61_4785_9de7_90b3ffaf0157.slice/crio-886b816e510ca0a1d185b2cae870d503dc001639fe916799eb14678839db73e5 WatchSource:0}: Error finding container 886b816e510ca0a1d185b2cae870d503dc001639fe916799eb14678839db73e5: Status 404 returned error can't find the container with id 886b816e510ca0a1d185b2cae870d503dc001639fe916799eb14678839db73e5 Nov 22 05:09:22 crc kubenswrapper[4818]: W1122 05:09:22.427085 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69f72526_235a_4079_a089_43e1b93aef9f.slice/crio-a508f74629c62988b53a3316a12fe97415911f0975215fa692d580fc2c3848c8 WatchSource:0}: Error finding container a508f74629c62988b53a3316a12fe97415911f0975215fa692d580fc2c3848c8: Status 404 returned error can't find the container with id a508f74629c62988b53a3316a12fe97415911f0975215fa692d580fc2c3848c8 Nov 22 05:09:22 crc kubenswrapper[4818]: I1122 05:09:22.431810 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 05:09:22 crc kubenswrapper[4818]: I1122 05:09:22.440182 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 22 05:09:22 crc kubenswrapper[4818]: W1122 05:09:22.444456 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9bfecc94_1073_46f1_9ea1_378ddc74ecf6.slice/crio-02429acc7fa6e2ac6602e6589ed94a96c5073e4ae9298faf4999b6d567024323 WatchSource:0}: Error finding container 02429acc7fa6e2ac6602e6589ed94a96c5073e4ae9298faf4999b6d567024323: Status 404 returned error can't find the container with id 02429acc7fa6e2ac6602e6589ed94a96c5073e4ae9298faf4999b6d567024323 Nov 22 05:09:22 crc kubenswrapper[4818]: W1122 05:09:22.445351 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4714282b_39ba_4691_8fbd_c3e76120fbb5.slice/crio-1a66248f3bcf6131f46a609cd10482bb28e7db21b16e4148cb684fe708230929 WatchSource:0}: Error finding container 1a66248f3bcf6131f46a609cd10482bb28e7db21b16e4148cb684fe708230929: Status 404 returned error can't find the container with id 1a66248f3bcf6131f46a609cd10482bb28e7db21b16e4148cb684fe708230929 Nov 22 05:09:22 crc kubenswrapper[4818]: I1122 05:09:22.447108 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-7c9sg"] Nov 22 05:09:22 crc kubenswrapper[4818]: I1122 05:09:22.657965 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 22 05:09:22 crc kubenswrapper[4818]: W1122 05:09:22.661246 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52d9e522_f84d_4671_9981_d08b3d7e0bbf.slice/crio-2b5249282986324b4513375581b526aa37a8d69bccb367360d84abab868f2033 WatchSource:0}: Error finding container 2b5249282986324b4513375581b526aa37a8d69bccb367360d84abab868f2033: Status 404 returned error can't find the container with id 2b5249282986324b4513375581b526aa37a8d69bccb367360d84abab868f2033 Nov 22 05:09:22 crc kubenswrapper[4818]: I1122 05:09:22.763102 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 22 05:09:22 crc kubenswrapper[4818]: W1122 05:09:22.764784 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72cac26c_ba82_41c7_a405_a9162867abea.slice/crio-6ddf6f2733d017b1b7e6d407842a2af1a547174e69bec1f4139130c1bfcfd002 WatchSource:0}: Error finding container 6ddf6f2733d017b1b7e6d407842a2af1a547174e69bec1f4139130c1bfcfd002: Status 404 returned error can't find the container with id 6ddf6f2733d017b1b7e6d407842a2af1a547174e69bec1f4139130c1bfcfd002 Nov 22 05:09:22 crc kubenswrapper[4818]: I1122 05:09:22.776600 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jgq5d"] Nov 22 05:09:22 crc kubenswrapper[4818]: I1122 05:09:22.795055 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 22 05:09:22 crc kubenswrapper[4818]: I1122 05:09:22.801215 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 05:09:22 crc kubenswrapper[4818]: W1122 05:09:22.806930 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b0746b7_8400_44c8_b4c5_c6b6e3811e91.slice/crio-658849f9eed0a62b8e8cfdc0bd84da447198facb64379d4b27f5cf065b668142 WatchSource:0}: Error finding container 658849f9eed0a62b8e8cfdc0bd84da447198facb64379d4b27f5cf065b668142: Status 404 returned error can't find the container with id 658849f9eed0a62b8e8cfdc0bd84da447198facb64379d4b27f5cf065b668142 Nov 22 05:09:22 crc kubenswrapper[4818]: I1122 05:09:22.809796 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 22 05:09:23 crc kubenswrapper[4818]: I1122 05:09:23.202725 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e98d9180-5e2b-4b38-8fc0-844072603c2d","Type":"ContainerStarted","Data":"4e7d496d26967286b087eb077dc85ffa5fdfb4fb401e33d50dd9b387c55db9eb"} Nov 22 05:09:23 crc kubenswrapper[4818]: I1122 05:09:23.204213 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"92a5743a-8fbe-49a7-a653-2b32e6034f52","Type":"ContainerStarted","Data":"8188a18c53b1e138d4155e4ec3e290ac468eb920c4360d5f90bbf0a1bf6e1b7c"} Nov 22 05:09:23 crc kubenswrapper[4818]: I1122 05:09:23.205486 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" event={"ID":"9bfecc94-1073-46f1-9ea1-378ddc74ecf6","Type":"ContainerStarted","Data":"02429acc7fa6e2ac6602e6589ed94a96c5073e4ae9298faf4999b6d567024323"} Nov 22 05:09:23 crc kubenswrapper[4818]: I1122 05:09:23.207018 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5b0746b7-8400-44c8-b4c5-c6b6e3811e91","Type":"ContainerStarted","Data":"658849f9eed0a62b8e8cfdc0bd84da447198facb64379d4b27f5cf065b668142"} Nov 22 05:09:23 crc kubenswrapper[4818]: I1122 05:09:23.209227 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"67e09220-fd61-4785-9de7-90b3ffaf0157","Type":"ContainerStarted","Data":"886b816e510ca0a1d185b2cae870d503dc001639fe916799eb14678839db73e5"} Nov 22 05:09:23 crc kubenswrapper[4818]: I1122 05:09:23.212736 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jgq5d" event={"ID":"7dec9ecf-e438-4334-a20e-757ce6fff445","Type":"ContainerStarted","Data":"5592ece4424087aafad0e052ad4c16c6294776b893464820e3c8cb01d11b2658"} Nov 22 05:09:23 crc kubenswrapper[4818]: I1122 05:09:23.214395 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"52d9e522-f84d-4671-9981-d08b3d7e0bbf","Type":"ContainerStarted","Data":"2b5249282986324b4513375581b526aa37a8d69bccb367360d84abab868f2033"} Nov 22 05:09:23 crc kubenswrapper[4818]: I1122 05:09:23.218802 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"72cac26c-ba82-41c7-a405-a9162867abea","Type":"ContainerStarted","Data":"6ddf6f2733d017b1b7e6d407842a2af1a547174e69bec1f4139130c1bfcfd002"} Nov 22 05:09:23 crc kubenswrapper[4818]: I1122 05:09:23.220247 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"4714282b-39ba-4691-8fbd-c3e76120fbb5","Type":"ContainerStarted","Data":"1a66248f3bcf6131f46a609cd10482bb28e7db21b16e4148cb684fe708230929"} Nov 22 05:09:23 crc kubenswrapper[4818]: I1122 05:09:23.221611 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"69f72526-235a-4079-a089-43e1b93aef9f","Type":"ContainerStarted","Data":"a508f74629c62988b53a3316a12fe97415911f0975215fa692d580fc2c3848c8"} Nov 22 05:09:23 crc kubenswrapper[4818]: I1122 05:09:23.801937 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-kpdh5"] Nov 22 05:09:23 crc kubenswrapper[4818]: E1122 05:09:23.989616 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 22 05:09:23 crc kubenswrapper[4818]: E1122 05:09:23.990403 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s8p48,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-nztvx_openstack(ee2898b0-81b0-473b-959d-e19bcc3836cf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 05:09:23 crc kubenswrapper[4818]: E1122 05:09:23.992624 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-nztvx" podUID="ee2898b0-81b0-473b-959d-e19bcc3836cf" Nov 22 05:09:24 crc kubenswrapper[4818]: E1122 05:09:24.034302 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 22 05:09:24 crc kubenswrapper[4818]: E1122 05:09:24.035146 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lk5ls,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-t62q2_openstack(f67a1621-8be8-490f-a2a5-489e11ba4b9e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 05:09:24 crc kubenswrapper[4818]: E1122 05:09:24.038376 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-t62q2" podUID="f67a1621-8be8-490f-a2a5-489e11ba4b9e" Nov 22 05:09:24 crc kubenswrapper[4818]: I1122 05:09:24.233756 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kpdh5" event={"ID":"1b31a0d3-3489-430a-853a-40d8715f7e7f","Type":"ContainerStarted","Data":"f820df94227fef1b8bf71cdcb91e3affc04d31b5321d11a7f89e3a13fa14349e"} Nov 22 05:09:24 crc kubenswrapper[4818]: I1122 05:09:24.656984 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-t62q2" Nov 22 05:09:24 crc kubenswrapper[4818]: I1122 05:09:24.664963 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-nztvx" Nov 22 05:09:24 crc kubenswrapper[4818]: I1122 05:09:24.820023 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8p48\" (UniqueName: \"kubernetes.io/projected/ee2898b0-81b0-473b-959d-e19bcc3836cf-kube-api-access-s8p48\") pod \"ee2898b0-81b0-473b-959d-e19bcc3836cf\" (UID: \"ee2898b0-81b0-473b-959d-e19bcc3836cf\") " Nov 22 05:09:24 crc kubenswrapper[4818]: I1122 05:09:24.820219 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk5ls\" (UniqueName: \"kubernetes.io/projected/f67a1621-8be8-490f-a2a5-489e11ba4b9e-kube-api-access-lk5ls\") pod \"f67a1621-8be8-490f-a2a5-489e11ba4b9e\" (UID: \"f67a1621-8be8-490f-a2a5-489e11ba4b9e\") " Nov 22 05:09:24 crc kubenswrapper[4818]: I1122 05:09:24.820300 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f67a1621-8be8-490f-a2a5-489e11ba4b9e-config\") pod \"f67a1621-8be8-490f-a2a5-489e11ba4b9e\" (UID: \"f67a1621-8be8-490f-a2a5-489e11ba4b9e\") " Nov 22 05:09:24 crc kubenswrapper[4818]: I1122 05:09:24.820377 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee2898b0-81b0-473b-959d-e19bcc3836cf-config\") pod \"ee2898b0-81b0-473b-959d-e19bcc3836cf\" (UID: \"ee2898b0-81b0-473b-959d-e19bcc3836cf\") " Nov 22 05:09:24 crc kubenswrapper[4818]: I1122 05:09:24.820440 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f67a1621-8be8-490f-a2a5-489e11ba4b9e-dns-svc\") pod \"f67a1621-8be8-490f-a2a5-489e11ba4b9e\" (UID: \"f67a1621-8be8-490f-a2a5-489e11ba4b9e\") " Nov 22 05:09:24 crc kubenswrapper[4818]: I1122 05:09:24.821159 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee2898b0-81b0-473b-959d-e19bcc3836cf-config" (OuterVolumeSpecName: "config") pod "ee2898b0-81b0-473b-959d-e19bcc3836cf" (UID: "ee2898b0-81b0-473b-959d-e19bcc3836cf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:09:24 crc kubenswrapper[4818]: I1122 05:09:24.821297 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f67a1621-8be8-490f-a2a5-489e11ba4b9e-config" (OuterVolumeSpecName: "config") pod "f67a1621-8be8-490f-a2a5-489e11ba4b9e" (UID: "f67a1621-8be8-490f-a2a5-489e11ba4b9e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:09:24 crc kubenswrapper[4818]: I1122 05:09:24.821387 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f67a1621-8be8-490f-a2a5-489e11ba4b9e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f67a1621-8be8-490f-a2a5-489e11ba4b9e" (UID: "f67a1621-8be8-490f-a2a5-489e11ba4b9e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:09:24 crc kubenswrapper[4818]: I1122 05:09:24.822097 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f67a1621-8be8-490f-a2a5-489e11ba4b9e-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:24 crc kubenswrapper[4818]: I1122 05:09:24.822131 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee2898b0-81b0-473b-959d-e19bcc3836cf-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:24 crc kubenswrapper[4818]: I1122 05:09:24.822151 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f67a1621-8be8-490f-a2a5-489e11ba4b9e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:24 crc kubenswrapper[4818]: I1122 05:09:24.827744 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f67a1621-8be8-490f-a2a5-489e11ba4b9e-kube-api-access-lk5ls" (OuterVolumeSpecName: "kube-api-access-lk5ls") pod "f67a1621-8be8-490f-a2a5-489e11ba4b9e" (UID: "f67a1621-8be8-490f-a2a5-489e11ba4b9e"). InnerVolumeSpecName "kube-api-access-lk5ls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:09:24 crc kubenswrapper[4818]: I1122 05:09:24.828647 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee2898b0-81b0-473b-959d-e19bcc3836cf-kube-api-access-s8p48" (OuterVolumeSpecName: "kube-api-access-s8p48") pod "ee2898b0-81b0-473b-959d-e19bcc3836cf" (UID: "ee2898b0-81b0-473b-959d-e19bcc3836cf"). InnerVolumeSpecName "kube-api-access-s8p48". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:09:24 crc kubenswrapper[4818]: I1122 05:09:24.923739 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk5ls\" (UniqueName: \"kubernetes.io/projected/f67a1621-8be8-490f-a2a5-489e11ba4b9e-kube-api-access-lk5ls\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:24 crc kubenswrapper[4818]: I1122 05:09:24.923775 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8p48\" (UniqueName: \"kubernetes.io/projected/ee2898b0-81b0-473b-959d-e19bcc3836cf-kube-api-access-s8p48\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:25 crc kubenswrapper[4818]: I1122 05:09:25.241003 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-nztvx" Nov 22 05:09:25 crc kubenswrapper[4818]: I1122 05:09:25.241001 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-nztvx" event={"ID":"ee2898b0-81b0-473b-959d-e19bcc3836cf","Type":"ContainerDied","Data":"8a653a934cc385cfd6eb4e907c0c02baa9bc5a73569e469d943df9429fde3a02"} Nov 22 05:09:25 crc kubenswrapper[4818]: I1122 05:09:25.245401 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-t62q2" event={"ID":"f67a1621-8be8-490f-a2a5-489e11ba4b9e","Type":"ContainerDied","Data":"89dcf2ffdb36837345685292f59d27013b3402e6166212c021a8952bcc32fc85"} Nov 22 05:09:25 crc kubenswrapper[4818]: I1122 05:09:25.245509 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-t62q2" Nov 22 05:09:25 crc kubenswrapper[4818]: I1122 05:09:25.319816 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nztvx"] Nov 22 05:09:25 crc kubenswrapper[4818]: I1122 05:09:25.327052 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nztvx"] Nov 22 05:09:25 crc kubenswrapper[4818]: I1122 05:09:25.339186 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-t62q2"] Nov 22 05:09:25 crc kubenswrapper[4818]: I1122 05:09:25.342348 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-t62q2"] Nov 22 05:09:26 crc kubenswrapper[4818]: I1122 05:09:26.257375 4818 generic.go:334] "Generic (PLEG): container finished" podID="9bfecc94-1073-46f1-9ea1-378ddc74ecf6" containerID="5d5306b64e468231f532921a551b1de29538680b45e8c132f45f3c1c317b7a23" exitCode=0 Nov 22 05:09:26 crc kubenswrapper[4818]: I1122 05:09:26.257425 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" event={"ID":"9bfecc94-1073-46f1-9ea1-378ddc74ecf6","Type":"ContainerDied","Data":"5d5306b64e468231f532921a551b1de29538680b45e8c132f45f3c1c317b7a23"} Nov 22 05:09:26 crc kubenswrapper[4818]: I1122 05:09:26.303646 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee2898b0-81b0-473b-959d-e19bcc3836cf" path="/var/lib/kubelet/pods/ee2898b0-81b0-473b-959d-e19bcc3836cf/volumes" Nov 22 05:09:26 crc kubenswrapper[4818]: I1122 05:09:26.304224 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f67a1621-8be8-490f-a2a5-489e11ba4b9e" path="/var/lib/kubelet/pods/f67a1621-8be8-490f-a2a5-489e11ba4b9e/volumes" Nov 22 05:09:33 crc kubenswrapper[4818]: I1122 05:09:33.314476 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" event={"ID":"9bfecc94-1073-46f1-9ea1-378ddc74ecf6","Type":"ContainerStarted","Data":"55631369b2274792243413de38c5e706cd22b3e8ec0eb4c68448c7d3d4949f91"} Nov 22 05:09:33 crc kubenswrapper[4818]: I1122 05:09:33.315363 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" Nov 22 05:09:33 crc kubenswrapper[4818]: I1122 05:09:33.337574 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" podStartSLOduration=35.661861421 podStartE2EDuration="38.33755061s" podCreationTimestamp="2025-11-22 05:08:55 +0000 UTC" firstStartedPulling="2025-11-22 05:09:22.44657387 +0000 UTC m=+1315.020990417" lastFinishedPulling="2025-11-22 05:09:25.122263079 +0000 UTC m=+1317.696679606" observedRunningTime="2025-11-22 05:09:33.335773412 +0000 UTC m=+1325.910189969" watchObservedRunningTime="2025-11-22 05:09:33.33755061 +0000 UTC m=+1325.911967167" Nov 22 05:09:34 crc kubenswrapper[4818]: I1122 05:09:34.334326 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5b0746b7-8400-44c8-b4c5-c6b6e3811e91","Type":"ContainerStarted","Data":"ab27ef86215761386ff54b32335f62fb4a3b1db3df58488f91ca5892ef99ed33"} Nov 22 05:09:34 crc kubenswrapper[4818]: I1122 05:09:34.336110 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"69f72526-235a-4079-a089-43e1b93aef9f","Type":"ContainerStarted","Data":"9f4ec022a951e769e06c755c5e3ee29d2cdb4bf72c3e18edaf341e5d425f9888"} Nov 22 05:09:34 crc kubenswrapper[4818]: I1122 05:09:34.338942 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kpdh5" event={"ID":"1b31a0d3-3489-430a-853a-40d8715f7e7f","Type":"ContainerStarted","Data":"aa3247e6e1ade1af38cc787be905da566d345d431af2dadc84059980dc6a3559"} Nov 22 05:09:34 crc kubenswrapper[4818]: I1122 05:09:34.342592 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jgq5d" event={"ID":"7dec9ecf-e438-4334-a20e-757ce6fff445","Type":"ContainerStarted","Data":"03b35807da2f0f657f8532c24c0116a2d4b5275f7a4f2dae04cf1751bb890d67"} Nov 22 05:09:34 crc kubenswrapper[4818]: I1122 05:09:34.342661 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-jgq5d" Nov 22 05:09:34 crc kubenswrapper[4818]: I1122 05:09:34.344046 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"4714282b-39ba-4691-8fbd-c3e76120fbb5","Type":"ContainerStarted","Data":"9bf08dd26af8a54e2b203a31efe61fa97578c7890450aaba61641dbf66849f3e"} Nov 22 05:09:34 crc kubenswrapper[4818]: I1122 05:09:34.346235 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e98d9180-5e2b-4b38-8fc0-844072603c2d","Type":"ContainerStarted","Data":"c32508a4202e11887410798e3c9721d3599cf6227a74984a9fb0d3fc6b35e5aa"} Nov 22 05:09:34 crc kubenswrapper[4818]: I1122 05:09:34.346466 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 22 05:09:34 crc kubenswrapper[4818]: I1122 05:09:34.347629 4818 generic.go:334] "Generic (PLEG): container finished" podID="acd97d29-b853-4de1-87e6-54d68a856eeb" containerID="ee229f1b4318aac332edff3651e2c3661077a0abbb789cfb13885cd26e9f4f9b" exitCode=0 Nov 22 05:09:34 crc kubenswrapper[4818]: I1122 05:09:34.347717 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" event={"ID":"acd97d29-b853-4de1-87e6-54d68a856eeb","Type":"ContainerDied","Data":"ee229f1b4318aac332edff3651e2c3661077a0abbb789cfb13885cd26e9f4f9b"} Nov 22 05:09:34 crc kubenswrapper[4818]: I1122 05:09:34.349519 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"52d9e522-f84d-4671-9981-d08b3d7e0bbf","Type":"ContainerStarted","Data":"346382f1a6e9f9ce713a0edca9f05ace8e7a433b4effdedb99ed4131c404c23c"} Nov 22 05:09:34 crc kubenswrapper[4818]: I1122 05:09:34.351546 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"72cac26c-ba82-41c7-a405-a9162867abea","Type":"ContainerStarted","Data":"960fee5ed9142bf83d1fc439eb20db601d18e38d15315f3239ca4a551e28ed7e"} Nov 22 05:09:34 crc kubenswrapper[4818]: I1122 05:09:34.352936 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"92a5743a-8fbe-49a7-a653-2b32e6034f52","Type":"ContainerStarted","Data":"705def928e920052d7ab5b829dc7e1b212eeb3dcaf5b1e7ed12249a5873c659b"} Nov 22 05:09:34 crc kubenswrapper[4818]: I1122 05:09:34.353298 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 22 05:09:34 crc kubenswrapper[4818]: I1122 05:09:34.354675 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"67e09220-fd61-4785-9de7-90b3ffaf0157","Type":"ContainerStarted","Data":"7b521d1554430e15d45e81217bcbedffcbf9f91f255aee93a639d50189365d64"} Nov 22 05:09:34 crc kubenswrapper[4818]: I1122 05:09:34.478210 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-jgq5d" podStartSLOduration=19.003238865 podStartE2EDuration="28.478188883s" podCreationTimestamp="2025-11-22 05:09:06 +0000 UTC" firstStartedPulling="2025-11-22 05:09:22.776155789 +0000 UTC m=+1315.350572326" lastFinishedPulling="2025-11-22 05:09:32.251105817 +0000 UTC m=+1324.825522344" observedRunningTime="2025-11-22 05:09:34.474933467 +0000 UTC m=+1327.049349994" watchObservedRunningTime="2025-11-22 05:09:34.478188883 +0000 UTC m=+1327.052605410" Nov 22 05:09:34 crc kubenswrapper[4818]: I1122 05:09:34.493817 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=26.489186371 podStartE2EDuration="35.493797838s" podCreationTimestamp="2025-11-22 05:08:59 +0000 UTC" firstStartedPulling="2025-11-22 05:09:22.786539955 +0000 UTC m=+1315.360956482" lastFinishedPulling="2025-11-22 05:09:31.791151422 +0000 UTC m=+1324.365567949" observedRunningTime="2025-11-22 05:09:34.490774088 +0000 UTC m=+1327.065190615" watchObservedRunningTime="2025-11-22 05:09:34.493797838 +0000 UTC m=+1327.068214365" Nov 22 05:09:34 crc kubenswrapper[4818]: I1122 05:09:34.528322 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=23.605479435 podStartE2EDuration="33.528295845s" podCreationTimestamp="2025-11-22 05:09:01 +0000 UTC" firstStartedPulling="2025-11-22 05:09:22.787053408 +0000 UTC m=+1315.361469935" lastFinishedPulling="2025-11-22 05:09:32.709869818 +0000 UTC m=+1325.284286345" observedRunningTime="2025-11-22 05:09:34.525945112 +0000 UTC m=+1327.100361639" watchObservedRunningTime="2025-11-22 05:09:34.528295845 +0000 UTC m=+1327.102712372" Nov 22 05:09:35 crc kubenswrapper[4818]: I1122 05:09:35.365240 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" event={"ID":"acd97d29-b853-4de1-87e6-54d68a856eeb","Type":"ContainerStarted","Data":"f3500e11719a18878d4746bc1aa4f9238ae0779d82545ce0b485d3aed621a6d2"} Nov 22 05:09:35 crc kubenswrapper[4818]: I1122 05:09:35.365533 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" Nov 22 05:09:35 crc kubenswrapper[4818]: I1122 05:09:35.369166 4818 generic.go:334] "Generic (PLEG): container finished" podID="1b31a0d3-3489-430a-853a-40d8715f7e7f" containerID="aa3247e6e1ade1af38cc787be905da566d345d431af2dadc84059980dc6a3559" exitCode=0 Nov 22 05:09:35 crc kubenswrapper[4818]: I1122 05:09:35.369283 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kpdh5" event={"ID":"1b31a0d3-3489-430a-853a-40d8715f7e7f","Type":"ContainerDied","Data":"aa3247e6e1ade1af38cc787be905da566d345d431af2dadc84059980dc6a3559"} Nov 22 05:09:35 crc kubenswrapper[4818]: I1122 05:09:35.390418 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" podStartSLOduration=-9223371996.464375 podStartE2EDuration="40.390400887s" podCreationTimestamp="2025-11-22 05:08:55 +0000 UTC" firstStartedPulling="2025-11-22 05:08:58.983347498 +0000 UTC m=+1291.557764025" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:09:35.382583819 +0000 UTC m=+1327.957000366" watchObservedRunningTime="2025-11-22 05:09:35.390400887 +0000 UTC m=+1327.964817414" Nov 22 05:09:36 crc kubenswrapper[4818]: I1122 05:09:36.380869 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kpdh5" event={"ID":"1b31a0d3-3489-430a-853a-40d8715f7e7f","Type":"ContainerStarted","Data":"1dd261907ed3bfcb086f8e0ddd4e80bc0f73c8958038a1514230de9041326a28"} Nov 22 05:09:36 crc kubenswrapper[4818]: I1122 05:09:36.381479 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kpdh5" event={"ID":"1b31a0d3-3489-430a-853a-40d8715f7e7f","Type":"ContainerStarted","Data":"e6def39d68efaa6949da34961ffcd3938f1c75310427b5cf5ed3962286f90dd0"} Nov 22 05:09:36 crc kubenswrapper[4818]: I1122 05:09:36.381502 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:36 crc kubenswrapper[4818]: I1122 05:09:36.381514 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:09:36 crc kubenswrapper[4818]: I1122 05:09:36.404696 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-kpdh5" podStartSLOduration=22.364329111 podStartE2EDuration="30.404672622s" podCreationTimestamp="2025-11-22 05:09:06 +0000 UTC" firstStartedPulling="2025-11-22 05:09:23.812436739 +0000 UTC m=+1316.386853276" lastFinishedPulling="2025-11-22 05:09:31.85278026 +0000 UTC m=+1324.427196787" observedRunningTime="2025-11-22 05:09:36.400927783 +0000 UTC m=+1328.975344310" watchObservedRunningTime="2025-11-22 05:09:36.404672622 +0000 UTC m=+1328.979089179" Nov 22 05:09:37 crc kubenswrapper[4818]: E1122 05:09:37.717925 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72cac26c_ba82_41c7_a405_a9162867abea.slice/crio-conmon-960fee5ed9142bf83d1fc439eb20db601d18e38d15315f3239ca4a551e28ed7e.scope\": RecentStats: unable to find data in memory cache]" Nov 22 05:09:38 crc kubenswrapper[4818]: I1122 05:09:38.412405 4818 generic.go:334] "Generic (PLEG): container finished" podID="72cac26c-ba82-41c7-a405-a9162867abea" containerID="960fee5ed9142bf83d1fc439eb20db601d18e38d15315f3239ca4a551e28ed7e" exitCode=0 Nov 22 05:09:38 crc kubenswrapper[4818]: I1122 05:09:38.412611 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"72cac26c-ba82-41c7-a405-a9162867abea","Type":"ContainerDied","Data":"960fee5ed9142bf83d1fc439eb20db601d18e38d15315f3239ca4a551e28ed7e"} Nov 22 05:09:38 crc kubenswrapper[4818]: I1122 05:09:38.418835 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"52d9e522-f84d-4671-9981-d08b3d7e0bbf","Type":"ContainerStarted","Data":"7d16f0cf925c485da439b4258a1066e6405e6a40a700ee0dae1dc005feb7e837"} Nov 22 05:09:38 crc kubenswrapper[4818]: I1122 05:09:38.423565 4818 generic.go:334] "Generic (PLEG): container finished" podID="4714282b-39ba-4691-8fbd-c3e76120fbb5" containerID="9bf08dd26af8a54e2b203a31efe61fa97578c7890450aaba61641dbf66849f3e" exitCode=0 Nov 22 05:09:38 crc kubenswrapper[4818]: I1122 05:09:38.423685 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"4714282b-39ba-4691-8fbd-c3e76120fbb5","Type":"ContainerDied","Data":"9bf08dd26af8a54e2b203a31efe61fa97578c7890450aaba61641dbf66849f3e"} Nov 22 05:09:38 crc kubenswrapper[4818]: I1122 05:09:38.426413 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5b0746b7-8400-44c8-b4c5-c6b6e3811e91","Type":"ContainerStarted","Data":"1aa7421fee0ae3b745a8b65550d4c1df37d17ddecfe045b01bda3fdd08213a68"} Nov 22 05:09:38 crc kubenswrapper[4818]: I1122 05:09:38.483349 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=16.806472431 podStartE2EDuration="31.483326585s" podCreationTimestamp="2025-11-22 05:09:07 +0000 UTC" firstStartedPulling="2025-11-22 05:09:22.663491524 +0000 UTC m=+1315.237908051" lastFinishedPulling="2025-11-22 05:09:37.340345668 +0000 UTC m=+1329.914762205" observedRunningTime="2025-11-22 05:09:38.473056191 +0000 UTC m=+1331.047472728" watchObservedRunningTime="2025-11-22 05:09:38.483326585 +0000 UTC m=+1331.057743112" Nov 22 05:09:38 crc kubenswrapper[4818]: I1122 05:09:38.509629 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=16.96444159 podStartE2EDuration="31.509610544s" podCreationTimestamp="2025-11-22 05:09:07 +0000 UTC" firstStartedPulling="2025-11-22 05:09:22.808993161 +0000 UTC m=+1315.383409688" lastFinishedPulling="2025-11-22 05:09:37.354162125 +0000 UTC m=+1329.928578642" observedRunningTime="2025-11-22 05:09:38.50235696 +0000 UTC m=+1331.076773507" watchObservedRunningTime="2025-11-22 05:09:38.509610544 +0000 UTC m=+1331.084027071" Nov 22 05:09:38 crc kubenswrapper[4818]: I1122 05:09:38.809171 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:38 crc kubenswrapper[4818]: I1122 05:09:38.809333 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:38 crc kubenswrapper[4818]: I1122 05:09:38.847367 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:38 crc kubenswrapper[4818]: I1122 05:09:38.962358 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:38 crc kubenswrapper[4818]: I1122 05:09:38.962410 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.008285 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.441656 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"4714282b-39ba-4691-8fbd-c3e76120fbb5","Type":"ContainerStarted","Data":"86d057b6acb49bd6b63ae27773599483f15e76d9a69464eeb270d5e7bbeb9d1c"} Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.453497 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"72cac26c-ba82-41c7-a405-a9162867abea","Type":"ContainerStarted","Data":"e55d62ddb7945f5862bc401e08a205178d70df9143a80d6bde3b6b67a73790b6"} Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.491283 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=32.688611786 podStartE2EDuration="42.491228201s" podCreationTimestamp="2025-11-22 05:08:57 +0000 UTC" firstStartedPulling="2025-11-22 05:09:22.447899405 +0000 UTC m=+1315.022315932" lastFinishedPulling="2025-11-22 05:09:32.25051582 +0000 UTC m=+1324.824932347" observedRunningTime="2025-11-22 05:09:39.48480759 +0000 UTC m=+1332.059224177" watchObservedRunningTime="2025-11-22 05:09:39.491228201 +0000 UTC m=+1332.065644758" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.520104 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=32.294848537 podStartE2EDuration="41.520078267s" podCreationTimestamp="2025-11-22 05:08:58 +0000 UTC" firstStartedPulling="2025-11-22 05:09:22.767680114 +0000 UTC m=+1315.342096661" lastFinishedPulling="2025-11-22 05:09:31.992909854 +0000 UTC m=+1324.567326391" observedRunningTime="2025-11-22 05:09:39.510154623 +0000 UTC m=+1332.084571180" watchObservedRunningTime="2025-11-22 05:09:39.520078267 +0000 UTC m=+1332.094494804" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.541725 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.549896 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.744168 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-7c9sg"] Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.745555 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" podUID="9bfecc94-1073-46f1-9ea1-378ddc74ecf6" containerName="dnsmasq-dns" containerID="cri-o://55631369b2274792243413de38c5e706cd22b3e8ec0eb4c68448c7d3d4949f91" gracePeriod=10 Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.751354 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.788316 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-4zbcq"] Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.790022 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.794034 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.813799 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-rsg7p"] Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.815071 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-rsg7p" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.817239 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.818964 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-4zbcq"] Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.824103 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-rsg7p"] Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.895056 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a26353d-d3de-497e-a6b0-260ca361ca1d-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-4zbcq\" (UID: \"2a26353d-d3de-497e-a6b0-260ca361ca1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.895326 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a26353d-d3de-497e-a6b0-260ca361ca1d-config\") pod \"dnsmasq-dns-7fd796d7df-4zbcq\" (UID: \"2a26353d-d3de-497e-a6b0-260ca361ca1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.895365 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a26353d-d3de-497e-a6b0-260ca361ca1d-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-4zbcq\" (UID: \"2a26353d-d3de-497e-a6b0-260ca361ca1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.895407 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg8k9\" (UniqueName: \"kubernetes.io/projected/2a26353d-d3de-497e-a6b0-260ca361ca1d-kube-api-access-lg8k9\") pod \"dnsmasq-dns-7fd796d7df-4zbcq\" (UID: \"2a26353d-d3de-497e-a6b0-260ca361ca1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.921656 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-j8tk7"] Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.921856 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" podUID="acd97d29-b853-4de1-87e6-54d68a856eeb" containerName="dnsmasq-dns" containerID="cri-o://f3500e11719a18878d4746bc1aa4f9238ae0779d82545ce0b485d3aed621a6d2" gracePeriod=10 Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.932425 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.944226 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.945578 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.952950 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.953024 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.953946 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-wrf4x" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.954176 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.954195 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.954365 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.959553 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.996715 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5b90b727-2867-4825-a5c1-a3549da28378-ovs-rundir\") pod \"ovn-controller-metrics-rsg7p\" (UID: \"5b90b727-2867-4825-a5c1-a3549da28378\") " pod="openstack/ovn-controller-metrics-rsg7p" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.996765 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slc9g\" (UniqueName: \"kubernetes.io/projected/5b90b727-2867-4825-a5c1-a3549da28378-kube-api-access-slc9g\") pod \"ovn-controller-metrics-rsg7p\" (UID: \"5b90b727-2867-4825-a5c1-a3549da28378\") " pod="openstack/ovn-controller-metrics-rsg7p" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.996819 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a26353d-d3de-497e-a6b0-260ca361ca1d-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-4zbcq\" (UID: \"2a26353d-d3de-497e-a6b0-260ca361ca1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.996846 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a26353d-d3de-497e-a6b0-260ca361ca1d-config\") pod \"dnsmasq-dns-7fd796d7df-4zbcq\" (UID: \"2a26353d-d3de-497e-a6b0-260ca361ca1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.996890 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a26353d-d3de-497e-a6b0-260ca361ca1d-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-4zbcq\" (UID: \"2a26353d-d3de-497e-a6b0-260ca361ca1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.996927 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b90b727-2867-4825-a5c1-a3549da28378-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-rsg7p\" (UID: \"5b90b727-2867-4825-a5c1-a3549da28378\") " pod="openstack/ovn-controller-metrics-rsg7p" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.996954 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5b90b727-2867-4825-a5c1-a3549da28378-ovn-rundir\") pod \"ovn-controller-metrics-rsg7p\" (UID: \"5b90b727-2867-4825-a5c1-a3549da28378\") " pod="openstack/ovn-controller-metrics-rsg7p" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.996989 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg8k9\" (UniqueName: \"kubernetes.io/projected/2a26353d-d3de-497e-a6b0-260ca361ca1d-kube-api-access-lg8k9\") pod \"dnsmasq-dns-7fd796d7df-4zbcq\" (UID: \"2a26353d-d3de-497e-a6b0-260ca361ca1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.997024 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b90b727-2867-4825-a5c1-a3549da28378-config\") pod \"ovn-controller-metrics-rsg7p\" (UID: \"5b90b727-2867-4825-a5c1-a3549da28378\") " pod="openstack/ovn-controller-metrics-rsg7p" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.997047 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b90b727-2867-4825-a5c1-a3549da28378-combined-ca-bundle\") pod \"ovn-controller-metrics-rsg7p\" (UID: \"5b90b727-2867-4825-a5c1-a3549da28378\") " pod="openstack/ovn-controller-metrics-rsg7p" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.998311 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a26353d-d3de-497e-a6b0-260ca361ca1d-config\") pod \"dnsmasq-dns-7fd796d7df-4zbcq\" (UID: \"2a26353d-d3de-497e-a6b0-260ca361ca1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.998695 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a26353d-d3de-497e-a6b0-260ca361ca1d-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-4zbcq\" (UID: \"2a26353d-d3de-497e-a6b0-260ca361ca1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" Nov 22 05:09:39 crc kubenswrapper[4818]: I1122 05:09:39.998827 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a26353d-d3de-497e-a6b0-260ca361ca1d-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-4zbcq\" (UID: \"2a26353d-d3de-497e-a6b0-260ca361ca1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.016806 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kfb5r"] Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.018066 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.026186 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.026775 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kfb5r"] Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.032980 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg8k9\" (UniqueName: \"kubernetes.io/projected/2a26353d-d3de-497e-a6b0-260ca361ca1d-kube-api-access-lg8k9\") pod \"dnsmasq-dns-7fd796d7df-4zbcq\" (UID: \"2a26353d-d3de-497e-a6b0-260ca361ca1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.098951 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5b90b727-2867-4825-a5c1-a3549da28378-ovs-rundir\") pod \"ovn-controller-metrics-rsg7p\" (UID: \"5b90b727-2867-4825-a5c1-a3549da28378\") " pod="openstack/ovn-controller-metrics-rsg7p" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.099001 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slc9g\" (UniqueName: \"kubernetes.io/projected/5b90b727-2867-4825-a5c1-a3549da28378-kube-api-access-slc9g\") pod \"ovn-controller-metrics-rsg7p\" (UID: \"5b90b727-2867-4825-a5c1-a3549da28378\") " pod="openstack/ovn-controller-metrics-rsg7p" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.099038 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-config\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.099059 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.099101 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b90b727-2867-4825-a5c1-a3549da28378-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-rsg7p\" (UID: \"5b90b727-2867-4825-a5c1-a3549da28378\") " pod="openstack/ovn-controller-metrics-rsg7p" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.099120 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5b90b727-2867-4825-a5c1-a3549da28378-ovn-rundir\") pod \"ovn-controller-metrics-rsg7p\" (UID: \"5b90b727-2867-4825-a5c1-a3549da28378\") " pod="openstack/ovn-controller-metrics-rsg7p" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.099145 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.099167 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkk4b\" (UniqueName: \"kubernetes.io/projected/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-kube-api-access-nkk4b\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.099183 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b90b727-2867-4825-a5c1-a3549da28378-config\") pod \"ovn-controller-metrics-rsg7p\" (UID: \"5b90b727-2867-4825-a5c1-a3549da28378\") " pod="openstack/ovn-controller-metrics-rsg7p" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.099199 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b90b727-2867-4825-a5c1-a3549da28378-combined-ca-bundle\") pod \"ovn-controller-metrics-rsg7p\" (UID: \"5b90b727-2867-4825-a5c1-a3549da28378\") " pod="openstack/ovn-controller-metrics-rsg7p" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.099217 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.099235 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.099282 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-scripts\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.099606 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5b90b727-2867-4825-a5c1-a3549da28378-ovn-rundir\") pod \"ovn-controller-metrics-rsg7p\" (UID: \"5b90b727-2867-4825-a5c1-a3549da28378\") " pod="openstack/ovn-controller-metrics-rsg7p" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.099688 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5b90b727-2867-4825-a5c1-a3549da28378-ovs-rundir\") pod \"ovn-controller-metrics-rsg7p\" (UID: \"5b90b727-2867-4825-a5c1-a3549da28378\") " pod="openstack/ovn-controller-metrics-rsg7p" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.100209 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b90b727-2867-4825-a5c1-a3549da28378-config\") pod \"ovn-controller-metrics-rsg7p\" (UID: \"5b90b727-2867-4825-a5c1-a3549da28378\") " pod="openstack/ovn-controller-metrics-rsg7p" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.103399 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b90b727-2867-4825-a5c1-a3549da28378-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-rsg7p\" (UID: \"5b90b727-2867-4825-a5c1-a3549da28378\") " pod="openstack/ovn-controller-metrics-rsg7p" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.103631 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b90b727-2867-4825-a5c1-a3549da28378-combined-ca-bundle\") pod \"ovn-controller-metrics-rsg7p\" (UID: \"5b90b727-2867-4825-a5c1-a3549da28378\") " pod="openstack/ovn-controller-metrics-rsg7p" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.120988 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.123552 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slc9g\" (UniqueName: \"kubernetes.io/projected/5b90b727-2867-4825-a5c1-a3549da28378-kube-api-access-slc9g\") pod \"ovn-controller-metrics-rsg7p\" (UID: \"5b90b727-2867-4825-a5c1-a3549da28378\") " pod="openstack/ovn-controller-metrics-rsg7p" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.202460 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-config\") pod \"dnsmasq-dns-86db49b7ff-kfb5r\" (UID: \"044b1787-72fc-46a3-b73b-fb412ebc4a28\") " pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.202509 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-kfb5r\" (UID: \"044b1787-72fc-46a3-b73b-fb412ebc4a28\") " pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.202533 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-kfb5r\" (UID: \"044b1787-72fc-46a3-b73b-fb412ebc4a28\") " pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.202570 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-kfb5r\" (UID: \"044b1787-72fc-46a3-b73b-fb412ebc4a28\") " pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.202589 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwfn8\" (UniqueName: \"kubernetes.io/projected/044b1787-72fc-46a3-b73b-fb412ebc4a28-kube-api-access-bwfn8\") pod \"dnsmasq-dns-86db49b7ff-kfb5r\" (UID: \"044b1787-72fc-46a3-b73b-fb412ebc4a28\") " pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.202617 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.202638 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkk4b\" (UniqueName: \"kubernetes.io/projected/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-kube-api-access-nkk4b\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.202663 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.202679 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.202708 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-scripts\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.202754 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-config\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.202774 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.204733 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.205341 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-scripts\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.205367 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-config\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.206474 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.209448 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.210170 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.224359 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkk4b\" (UniqueName: \"kubernetes.io/projected/6ab8b0e2-6df1-4a6b-b800-93e5b51ee430-kube-api-access-nkk4b\") pod \"ovn-northd-0\" (UID: \"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430\") " pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.240082 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-rsg7p" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.289007 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.304406 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-kfb5r\" (UID: \"044b1787-72fc-46a3-b73b-fb412ebc4a28\") " pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.304444 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-kfb5r\" (UID: \"044b1787-72fc-46a3-b73b-fb412ebc4a28\") " pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.304497 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-kfb5r\" (UID: \"044b1787-72fc-46a3-b73b-fb412ebc4a28\") " pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.304518 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwfn8\" (UniqueName: \"kubernetes.io/projected/044b1787-72fc-46a3-b73b-fb412ebc4a28-kube-api-access-bwfn8\") pod \"dnsmasq-dns-86db49b7ff-kfb5r\" (UID: \"044b1787-72fc-46a3-b73b-fb412ebc4a28\") " pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.304633 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-config\") pod \"dnsmasq-dns-86db49b7ff-kfb5r\" (UID: \"044b1787-72fc-46a3-b73b-fb412ebc4a28\") " pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.305478 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-kfb5r\" (UID: \"044b1787-72fc-46a3-b73b-fb412ebc4a28\") " pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.306048 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-kfb5r\" (UID: \"044b1787-72fc-46a3-b73b-fb412ebc4a28\") " pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.306123 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-kfb5r\" (UID: \"044b1787-72fc-46a3-b73b-fb412ebc4a28\") " pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.306793 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-config\") pod \"dnsmasq-dns-86db49b7ff-kfb5r\" (UID: \"044b1787-72fc-46a3-b73b-fb412ebc4a28\") " pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.327918 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.336160 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.336526 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwfn8\" (UniqueName: \"kubernetes.io/projected/044b1787-72fc-46a3-b73b-fb412ebc4a28-kube-api-access-bwfn8\") pod \"dnsmasq-dns-86db49b7ff-kfb5r\" (UID: \"044b1787-72fc-46a3-b73b-fb412ebc4a28\") " pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.406834 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvcnk\" (UniqueName: \"kubernetes.io/projected/9bfecc94-1073-46f1-9ea1-378ddc74ecf6-kube-api-access-rvcnk\") pod \"9bfecc94-1073-46f1-9ea1-378ddc74ecf6\" (UID: \"9bfecc94-1073-46f1-9ea1-378ddc74ecf6\") " Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.406942 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bfecc94-1073-46f1-9ea1-378ddc74ecf6-dns-svc\") pod \"9bfecc94-1073-46f1-9ea1-378ddc74ecf6\" (UID: \"9bfecc94-1073-46f1-9ea1-378ddc74ecf6\") " Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.406993 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bfecc94-1073-46f1-9ea1-378ddc74ecf6-config\") pod \"9bfecc94-1073-46f1-9ea1-378ddc74ecf6\" (UID: \"9bfecc94-1073-46f1-9ea1-378ddc74ecf6\") " Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.412139 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bfecc94-1073-46f1-9ea1-378ddc74ecf6-kube-api-access-rvcnk" (OuterVolumeSpecName: "kube-api-access-rvcnk") pod "9bfecc94-1073-46f1-9ea1-378ddc74ecf6" (UID: "9bfecc94-1073-46f1-9ea1-378ddc74ecf6"). InnerVolumeSpecName "kube-api-access-rvcnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.428551 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.465982 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.472175 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bfecc94-1073-46f1-9ea1-378ddc74ecf6-config" (OuterVolumeSpecName: "config") pod "9bfecc94-1073-46f1-9ea1-378ddc74ecf6" (UID: "9bfecc94-1073-46f1-9ea1-378ddc74ecf6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.473161 4818 generic.go:334] "Generic (PLEG): container finished" podID="acd97d29-b853-4de1-87e6-54d68a856eeb" containerID="f3500e11719a18878d4746bc1aa4f9238ae0779d82545ce0b485d3aed621a6d2" exitCode=0 Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.474604 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" event={"ID":"acd97d29-b853-4de1-87e6-54d68a856eeb","Type":"ContainerDied","Data":"f3500e11719a18878d4746bc1aa4f9238ae0779d82545ce0b485d3aed621a6d2"} Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.474709 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" event={"ID":"acd97d29-b853-4de1-87e6-54d68a856eeb","Type":"ContainerDied","Data":"7499d439594b45be495f421f475caa694a5e85b38a1f62a7416b1eeed941110c"} Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.474790 4818 scope.go:117] "RemoveContainer" containerID="f3500e11719a18878d4746bc1aa4f9238ae0779d82545ce0b485d3aed621a6d2" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.474998 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-j8tk7" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.480832 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bfecc94-1073-46f1-9ea1-378ddc74ecf6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9bfecc94-1073-46f1-9ea1-378ddc74ecf6" (UID: "9bfecc94-1073-46f1-9ea1-378ddc74ecf6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.489091 4818 generic.go:334] "Generic (PLEG): container finished" podID="9bfecc94-1073-46f1-9ea1-378ddc74ecf6" containerID="55631369b2274792243413de38c5e706cd22b3e8ec0eb4c68448c7d3d4949f91" exitCode=0 Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.489133 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" event={"ID":"9bfecc94-1073-46f1-9ea1-378ddc74ecf6","Type":"ContainerDied","Data":"55631369b2274792243413de38c5e706cd22b3e8ec0eb4c68448c7d3d4949f91"} Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.489172 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" event={"ID":"9bfecc94-1073-46f1-9ea1-378ddc74ecf6","Type":"ContainerDied","Data":"02429acc7fa6e2ac6602e6589ed94a96c5073e4ae9298faf4999b6d567024323"} Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.489169 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-7c9sg" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.509131 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvcnk\" (UniqueName: \"kubernetes.io/projected/9bfecc94-1073-46f1-9ea1-378ddc74ecf6-kube-api-access-rvcnk\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.509164 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bfecc94-1073-46f1-9ea1-378ddc74ecf6-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.509174 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bfecc94-1073-46f1-9ea1-378ddc74ecf6-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.533172 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-7c9sg"] Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.543523 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-7c9sg"] Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.543646 4818 scope.go:117] "RemoveContainer" containerID="ee229f1b4318aac332edff3651e2c3661077a0abbb789cfb13885cd26e9f4f9b" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.563797 4818 scope.go:117] "RemoveContainer" containerID="f3500e11719a18878d4746bc1aa4f9238ae0779d82545ce0b485d3aed621a6d2" Nov 22 05:09:40 crc kubenswrapper[4818]: E1122 05:09:40.564730 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3500e11719a18878d4746bc1aa4f9238ae0779d82545ce0b485d3aed621a6d2\": container with ID starting with f3500e11719a18878d4746bc1aa4f9238ae0779d82545ce0b485d3aed621a6d2 not found: ID does not exist" containerID="f3500e11719a18878d4746bc1aa4f9238ae0779d82545ce0b485d3aed621a6d2" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.564773 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3500e11719a18878d4746bc1aa4f9238ae0779d82545ce0b485d3aed621a6d2"} err="failed to get container status \"f3500e11719a18878d4746bc1aa4f9238ae0779d82545ce0b485d3aed621a6d2\": rpc error: code = NotFound desc = could not find container \"f3500e11719a18878d4746bc1aa4f9238ae0779d82545ce0b485d3aed621a6d2\": container with ID starting with f3500e11719a18878d4746bc1aa4f9238ae0779d82545ce0b485d3aed621a6d2 not found: ID does not exist" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.564791 4818 scope.go:117] "RemoveContainer" containerID="ee229f1b4318aac332edff3651e2c3661077a0abbb789cfb13885cd26e9f4f9b" Nov 22 05:09:40 crc kubenswrapper[4818]: E1122 05:09:40.566938 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee229f1b4318aac332edff3651e2c3661077a0abbb789cfb13885cd26e9f4f9b\": container with ID starting with ee229f1b4318aac332edff3651e2c3661077a0abbb789cfb13885cd26e9f4f9b not found: ID does not exist" containerID="ee229f1b4318aac332edff3651e2c3661077a0abbb789cfb13885cd26e9f4f9b" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.566967 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee229f1b4318aac332edff3651e2c3661077a0abbb789cfb13885cd26e9f4f9b"} err="failed to get container status \"ee229f1b4318aac332edff3651e2c3661077a0abbb789cfb13885cd26e9f4f9b\": rpc error: code = NotFound desc = could not find container \"ee229f1b4318aac332edff3651e2c3661077a0abbb789cfb13885cd26e9f4f9b\": container with ID starting with ee229f1b4318aac332edff3651e2c3661077a0abbb789cfb13885cd26e9f4f9b not found: ID does not exist" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.566983 4818 scope.go:117] "RemoveContainer" containerID="55631369b2274792243413de38c5e706cd22b3e8ec0eb4c68448c7d3d4949f91" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.588675 4818 scope.go:117] "RemoveContainer" containerID="5d5306b64e468231f532921a551b1de29538680b45e8c132f45f3c1c317b7a23" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.610219 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acd97d29-b853-4de1-87e6-54d68a856eeb-config\") pod \"acd97d29-b853-4de1-87e6-54d68a856eeb\" (UID: \"acd97d29-b853-4de1-87e6-54d68a856eeb\") " Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.610386 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdf7f\" (UniqueName: \"kubernetes.io/projected/acd97d29-b853-4de1-87e6-54d68a856eeb-kube-api-access-sdf7f\") pod \"acd97d29-b853-4de1-87e6-54d68a856eeb\" (UID: \"acd97d29-b853-4de1-87e6-54d68a856eeb\") " Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.610460 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/acd97d29-b853-4de1-87e6-54d68a856eeb-dns-svc\") pod \"acd97d29-b853-4de1-87e6-54d68a856eeb\" (UID: \"acd97d29-b853-4de1-87e6-54d68a856eeb\") " Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.615701 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acd97d29-b853-4de1-87e6-54d68a856eeb-kube-api-access-sdf7f" (OuterVolumeSpecName: "kube-api-access-sdf7f") pod "acd97d29-b853-4de1-87e6-54d68a856eeb" (UID: "acd97d29-b853-4de1-87e6-54d68a856eeb"). InnerVolumeSpecName "kube-api-access-sdf7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.623379 4818 scope.go:117] "RemoveContainer" containerID="55631369b2274792243413de38c5e706cd22b3e8ec0eb4c68448c7d3d4949f91" Nov 22 05:09:40 crc kubenswrapper[4818]: E1122 05:09:40.629469 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55631369b2274792243413de38c5e706cd22b3e8ec0eb4c68448c7d3d4949f91\": container with ID starting with 55631369b2274792243413de38c5e706cd22b3e8ec0eb4c68448c7d3d4949f91 not found: ID does not exist" containerID="55631369b2274792243413de38c5e706cd22b3e8ec0eb4c68448c7d3d4949f91" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.629614 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55631369b2274792243413de38c5e706cd22b3e8ec0eb4c68448c7d3d4949f91"} err="failed to get container status \"55631369b2274792243413de38c5e706cd22b3e8ec0eb4c68448c7d3d4949f91\": rpc error: code = NotFound desc = could not find container \"55631369b2274792243413de38c5e706cd22b3e8ec0eb4c68448c7d3d4949f91\": container with ID starting with 55631369b2274792243413de38c5e706cd22b3e8ec0eb4c68448c7d3d4949f91 not found: ID does not exist" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.629642 4818 scope.go:117] "RemoveContainer" containerID="5d5306b64e468231f532921a551b1de29538680b45e8c132f45f3c1c317b7a23" Nov 22 05:09:40 crc kubenswrapper[4818]: E1122 05:09:40.630286 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d5306b64e468231f532921a551b1de29538680b45e8c132f45f3c1c317b7a23\": container with ID starting with 5d5306b64e468231f532921a551b1de29538680b45e8c132f45f3c1c317b7a23 not found: ID does not exist" containerID="5d5306b64e468231f532921a551b1de29538680b45e8c132f45f3c1c317b7a23" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.630315 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d5306b64e468231f532921a551b1de29538680b45e8c132f45f3c1c317b7a23"} err="failed to get container status \"5d5306b64e468231f532921a551b1de29538680b45e8c132f45f3c1c317b7a23\": rpc error: code = NotFound desc = could not find container \"5d5306b64e468231f532921a551b1de29538680b45e8c132f45f3c1c317b7a23\": container with ID starting with 5d5306b64e468231f532921a551b1de29538680b45e8c132f45f3c1c317b7a23 not found: ID does not exist" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.648903 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-4zbcq"] Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.682021 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acd97d29-b853-4de1-87e6-54d68a856eeb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "acd97d29-b853-4de1-87e6-54d68a856eeb" (UID: "acd97d29-b853-4de1-87e6-54d68a856eeb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.695338 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acd97d29-b853-4de1-87e6-54d68a856eeb-config" (OuterVolumeSpecName: "config") pod "acd97d29-b853-4de1-87e6-54d68a856eeb" (UID: "acd97d29-b853-4de1-87e6-54d68a856eeb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.713591 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acd97d29-b853-4de1-87e6-54d68a856eeb-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.713616 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdf7f\" (UniqueName: \"kubernetes.io/projected/acd97d29-b853-4de1-87e6-54d68a856eeb-kube-api-access-sdf7f\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.713627 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/acd97d29-b853-4de1-87e6-54d68a856eeb-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.841598 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kfb5r"] Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.862390 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-rsg7p"] Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.868999 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-j8tk7"] Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.877407 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-j8tk7"] Nov 22 05:09:40 crc kubenswrapper[4818]: I1122 05:09:40.932073 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 22 05:09:40 crc kubenswrapper[4818]: W1122 05:09:40.935143 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ab8b0e2_6df1_4a6b_b800_93e5b51ee430.slice/crio-0f75e96281d456758195b969fd4384343ae4a623c87508c7394115823354dc1f WatchSource:0}: Error finding container 0f75e96281d456758195b969fd4384343ae4a623c87508c7394115823354dc1f: Status 404 returned error can't find the container with id 0f75e96281d456758195b969fd4384343ae4a623c87508c7394115823354dc1f Nov 22 05:09:41 crc kubenswrapper[4818]: I1122 05:09:41.500458 4818 generic.go:334] "Generic (PLEG): container finished" podID="044b1787-72fc-46a3-b73b-fb412ebc4a28" containerID="de1bf831a88a2bc5fd75be68335d24ca6ec6611149ccb8c5f88ca9b55dd3e3f4" exitCode=0 Nov 22 05:09:41 crc kubenswrapper[4818]: I1122 05:09:41.500630 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" event={"ID":"044b1787-72fc-46a3-b73b-fb412ebc4a28","Type":"ContainerDied","Data":"de1bf831a88a2bc5fd75be68335d24ca6ec6611149ccb8c5f88ca9b55dd3e3f4"} Nov 22 05:09:41 crc kubenswrapper[4818]: I1122 05:09:41.500802 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" event={"ID":"044b1787-72fc-46a3-b73b-fb412ebc4a28","Type":"ContainerStarted","Data":"f4589124ace1bb63071aa23a6df309ac67df4dda79dbd8776f65796a84f947d4"} Nov 22 05:09:41 crc kubenswrapper[4818]: I1122 05:09:41.506112 4818 generic.go:334] "Generic (PLEG): container finished" podID="2a26353d-d3de-497e-a6b0-260ca361ca1d" containerID="878ab50cf31c6349bdb71aea880bc54b3848abe72d9af7ad74c6e73510ab7788" exitCode=0 Nov 22 05:09:41 crc kubenswrapper[4818]: I1122 05:09:41.506227 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" event={"ID":"2a26353d-d3de-497e-a6b0-260ca361ca1d","Type":"ContainerDied","Data":"878ab50cf31c6349bdb71aea880bc54b3848abe72d9af7ad74c6e73510ab7788"} Nov 22 05:09:41 crc kubenswrapper[4818]: I1122 05:09:41.506294 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" event={"ID":"2a26353d-d3de-497e-a6b0-260ca361ca1d","Type":"ContainerStarted","Data":"cc64b8500b3d94f041e4830dc12232b54bbcf022cb5531ae70bff2e98af4fc11"} Nov 22 05:09:41 crc kubenswrapper[4818]: I1122 05:09:41.514912 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430","Type":"ContainerStarted","Data":"0f75e96281d456758195b969fd4384343ae4a623c87508c7394115823354dc1f"} Nov 22 05:09:41 crc kubenswrapper[4818]: I1122 05:09:41.517576 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-rsg7p" event={"ID":"5b90b727-2867-4825-a5c1-a3549da28378","Type":"ContainerStarted","Data":"eeac664e1b706749c9ef80cd7b465d96d03873872f6279fd5bd024327657c221"} Nov 22 05:09:41 crc kubenswrapper[4818]: I1122 05:09:41.517731 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-rsg7p" event={"ID":"5b90b727-2867-4825-a5c1-a3549da28378","Type":"ContainerStarted","Data":"84f6715af4d4f2f50fa6b1a9a571c4cf524a58c15371c154fed897a151b8e95f"} Nov 22 05:09:41 crc kubenswrapper[4818]: I1122 05:09:41.547999 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-rsg7p" podStartSLOduration=2.5479790209999997 podStartE2EDuration="2.547979021s" podCreationTimestamp="2025-11-22 05:09:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:09:41.538573741 +0000 UTC m=+1334.112990278" watchObservedRunningTime="2025-11-22 05:09:41.547979021 +0000 UTC m=+1334.122395558" Nov 22 05:09:41 crc kubenswrapper[4818]: I1122 05:09:41.952749 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 22 05:09:42 crc kubenswrapper[4818]: I1122 05:09:42.303732 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bfecc94-1073-46f1-9ea1-378ddc74ecf6" path="/var/lib/kubelet/pods/9bfecc94-1073-46f1-9ea1-378ddc74ecf6/volumes" Nov 22 05:09:42 crc kubenswrapper[4818]: I1122 05:09:42.304442 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acd97d29-b853-4de1-87e6-54d68a856eeb" path="/var/lib/kubelet/pods/acd97d29-b853-4de1-87e6-54d68a856eeb/volumes" Nov 22 05:09:42 crc kubenswrapper[4818]: I1122 05:09:42.527358 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" event={"ID":"044b1787-72fc-46a3-b73b-fb412ebc4a28","Type":"ContainerStarted","Data":"92f506c62a4fd3211edc2e9d37d390268922b279faaec8732889f20b9cdcdd78"} Nov 22 05:09:42 crc kubenswrapper[4818]: I1122 05:09:42.527684 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:09:42 crc kubenswrapper[4818]: I1122 05:09:42.534820 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" event={"ID":"2a26353d-d3de-497e-a6b0-260ca361ca1d","Type":"ContainerStarted","Data":"5d4a3c160d8959308cc0165218bb8933482204d3fa3f9e867d62f3fe3f93fc56"} Nov 22 05:09:42 crc kubenswrapper[4818]: I1122 05:09:42.534973 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" Nov 22 05:09:42 crc kubenswrapper[4818]: I1122 05:09:42.537224 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430","Type":"ContainerStarted","Data":"8b6d38fa881c00b00b2f461ee4f3b954eaa2aabd89640b9bb6f743d0eb2da791"} Nov 22 05:09:42 crc kubenswrapper[4818]: I1122 05:09:42.547421 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" podStartSLOduration=3.547402471 podStartE2EDuration="3.547402471s" podCreationTimestamp="2025-11-22 05:09:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:09:42.545204293 +0000 UTC m=+1335.119620840" watchObservedRunningTime="2025-11-22 05:09:42.547402471 +0000 UTC m=+1335.121818998" Nov 22 05:09:42 crc kubenswrapper[4818]: I1122 05:09:42.567334 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" podStartSLOduration=3.567312411 podStartE2EDuration="3.567312411s" podCreationTimestamp="2025-11-22 05:09:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:09:42.563289134 +0000 UTC m=+1335.137705671" watchObservedRunningTime="2025-11-22 05:09:42.567312411 +0000 UTC m=+1335.141728938" Nov 22 05:09:43 crc kubenswrapper[4818]: I1122 05:09:43.547147 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"6ab8b0e2-6df1-4a6b-b800-93e5b51ee430","Type":"ContainerStarted","Data":"590a35817090b647da9955539dda4601a41ca6a181ec47792994f5932f2bfc11"} Nov 22 05:09:43 crc kubenswrapper[4818]: I1122 05:09:43.583362 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.2914520290000002 podStartE2EDuration="4.583323612s" podCreationTimestamp="2025-11-22 05:09:39 +0000 UTC" firstStartedPulling="2025-11-22 05:09:40.938139274 +0000 UTC m=+1333.512555801" lastFinishedPulling="2025-11-22 05:09:42.230010857 +0000 UTC m=+1334.804427384" observedRunningTime="2025-11-22 05:09:43.572865765 +0000 UTC m=+1336.147282392" watchObservedRunningTime="2025-11-22 05:09:43.583323612 +0000 UTC m=+1336.157740189" Nov 22 05:09:44 crc kubenswrapper[4818]: I1122 05:09:44.556999 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 22 05:09:46 crc kubenswrapper[4818]: I1122 05:09:46.092223 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 22 05:09:46 crc kubenswrapper[4818]: I1122 05:09:46.192803 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 22 05:09:48 crc kubenswrapper[4818]: I1122 05:09:48.369651 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 22 05:09:48 crc kubenswrapper[4818]: I1122 05:09:48.370646 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 22 05:09:48 crc kubenswrapper[4818]: I1122 05:09:48.485074 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 22 05:09:48 crc kubenswrapper[4818]: I1122 05:09:48.705038 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 22 05:09:49 crc kubenswrapper[4818]: I1122 05:09:49.992561 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-4fc8-account-create-4nkhn"] Nov 22 05:09:49 crc kubenswrapper[4818]: E1122 05:09:49.992956 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bfecc94-1073-46f1-9ea1-378ddc74ecf6" containerName="dnsmasq-dns" Nov 22 05:09:49 crc kubenswrapper[4818]: I1122 05:09:49.992971 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bfecc94-1073-46f1-9ea1-378ddc74ecf6" containerName="dnsmasq-dns" Nov 22 05:09:49 crc kubenswrapper[4818]: E1122 05:09:49.992984 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bfecc94-1073-46f1-9ea1-378ddc74ecf6" containerName="init" Nov 22 05:09:49 crc kubenswrapper[4818]: I1122 05:09:49.992990 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bfecc94-1073-46f1-9ea1-378ddc74ecf6" containerName="init" Nov 22 05:09:49 crc kubenswrapper[4818]: E1122 05:09:49.993004 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acd97d29-b853-4de1-87e6-54d68a856eeb" containerName="dnsmasq-dns" Nov 22 05:09:49 crc kubenswrapper[4818]: I1122 05:09:49.993010 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="acd97d29-b853-4de1-87e6-54d68a856eeb" containerName="dnsmasq-dns" Nov 22 05:09:49 crc kubenswrapper[4818]: E1122 05:09:49.993024 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acd97d29-b853-4de1-87e6-54d68a856eeb" containerName="init" Nov 22 05:09:49 crc kubenswrapper[4818]: I1122 05:09:49.993031 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="acd97d29-b853-4de1-87e6-54d68a856eeb" containerName="init" Nov 22 05:09:49 crc kubenswrapper[4818]: I1122 05:09:49.993170 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bfecc94-1073-46f1-9ea1-378ddc74ecf6" containerName="dnsmasq-dns" Nov 22 05:09:49 crc kubenswrapper[4818]: I1122 05:09:49.993185 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="acd97d29-b853-4de1-87e6-54d68a856eeb" containerName="dnsmasq-dns" Nov 22 05:09:49 crc kubenswrapper[4818]: I1122 05:09:49.993752 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4fc8-account-create-4nkhn" Nov 22 05:09:49 crc kubenswrapper[4818]: I1122 05:09:49.996045 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.003043 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-4fc8-account-create-4nkhn"] Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.058300 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-gmlh9"] Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.059509 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gmlh9" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.067334 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gmlh9"] Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.079836 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9hxz\" (UniqueName: \"kubernetes.io/projected/5b0bbbc5-6ad1-4390-8d2b-b91c83f49701-kube-api-access-w9hxz\") pod \"keystone-4fc8-account-create-4nkhn\" (UID: \"5b0bbbc5-6ad1-4390-8d2b-b91c83f49701\") " pod="openstack/keystone-4fc8-account-create-4nkhn" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.079948 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b0bbbc5-6ad1-4390-8d2b-b91c83f49701-operator-scripts\") pod \"keystone-4fc8-account-create-4nkhn\" (UID: \"5b0bbbc5-6ad1-4390-8d2b-b91c83f49701\") " pod="openstack/keystone-4fc8-account-create-4nkhn" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.079990 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt9pd\" (UniqueName: \"kubernetes.io/projected/4f234b24-4e7a-42dc-943d-aa45cb9cd72b-kube-api-access-dt9pd\") pod \"keystone-db-create-gmlh9\" (UID: \"4f234b24-4e7a-42dc-943d-aa45cb9cd72b\") " pod="openstack/keystone-db-create-gmlh9" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.080119 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f234b24-4e7a-42dc-943d-aa45cb9cd72b-operator-scripts\") pod \"keystone-db-create-gmlh9\" (UID: \"4f234b24-4e7a-42dc-943d-aa45cb9cd72b\") " pod="openstack/keystone-db-create-gmlh9" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.122433 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.182383 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9hxz\" (UniqueName: \"kubernetes.io/projected/5b0bbbc5-6ad1-4390-8d2b-b91c83f49701-kube-api-access-w9hxz\") pod \"keystone-4fc8-account-create-4nkhn\" (UID: \"5b0bbbc5-6ad1-4390-8d2b-b91c83f49701\") " pod="openstack/keystone-4fc8-account-create-4nkhn" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.182475 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b0bbbc5-6ad1-4390-8d2b-b91c83f49701-operator-scripts\") pod \"keystone-4fc8-account-create-4nkhn\" (UID: \"5b0bbbc5-6ad1-4390-8d2b-b91c83f49701\") " pod="openstack/keystone-4fc8-account-create-4nkhn" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.182513 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt9pd\" (UniqueName: \"kubernetes.io/projected/4f234b24-4e7a-42dc-943d-aa45cb9cd72b-kube-api-access-dt9pd\") pod \"keystone-db-create-gmlh9\" (UID: \"4f234b24-4e7a-42dc-943d-aa45cb9cd72b\") " pod="openstack/keystone-db-create-gmlh9" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.182550 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f234b24-4e7a-42dc-943d-aa45cb9cd72b-operator-scripts\") pod \"keystone-db-create-gmlh9\" (UID: \"4f234b24-4e7a-42dc-943d-aa45cb9cd72b\") " pod="openstack/keystone-db-create-gmlh9" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.183412 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f234b24-4e7a-42dc-943d-aa45cb9cd72b-operator-scripts\") pod \"keystone-db-create-gmlh9\" (UID: \"4f234b24-4e7a-42dc-943d-aa45cb9cd72b\") " pod="openstack/keystone-db-create-gmlh9" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.183604 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b0bbbc5-6ad1-4390-8d2b-b91c83f49701-operator-scripts\") pod \"keystone-4fc8-account-create-4nkhn\" (UID: \"5b0bbbc5-6ad1-4390-8d2b-b91c83f49701\") " pod="openstack/keystone-4fc8-account-create-4nkhn" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.213685 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9hxz\" (UniqueName: \"kubernetes.io/projected/5b0bbbc5-6ad1-4390-8d2b-b91c83f49701-kube-api-access-w9hxz\") pod \"keystone-4fc8-account-create-4nkhn\" (UID: \"5b0bbbc5-6ad1-4390-8d2b-b91c83f49701\") " pod="openstack/keystone-4fc8-account-create-4nkhn" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.234068 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt9pd\" (UniqueName: \"kubernetes.io/projected/4f234b24-4e7a-42dc-943d-aa45cb9cd72b-kube-api-access-dt9pd\") pod \"keystone-db-create-gmlh9\" (UID: \"4f234b24-4e7a-42dc-943d-aa45cb9cd72b\") " pod="openstack/keystone-db-create-gmlh9" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.260240 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-wg496"] Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.261541 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wg496" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.270316 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-wg496"] Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.283833 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqdvk\" (UniqueName: \"kubernetes.io/projected/501b7faf-95bc-41d8-8b80-20c1c144395a-kube-api-access-bqdvk\") pod \"placement-db-create-wg496\" (UID: \"501b7faf-95bc-41d8-8b80-20c1c144395a\") " pod="openstack/placement-db-create-wg496" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.283898 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/501b7faf-95bc-41d8-8b80-20c1c144395a-operator-scripts\") pod \"placement-db-create-wg496\" (UID: \"501b7faf-95bc-41d8-8b80-20c1c144395a\") " pod="openstack/placement-db-create-wg496" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.313398 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4fc8-account-create-4nkhn" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.350985 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-b756-account-create-n4p6g"] Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.355221 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b756-account-create-n4p6g" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.366656 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.378454 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gmlh9" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.383185 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-b756-account-create-n4p6g"] Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.386214 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff776fa9-dc41-4535-8a27-579d82db15b1-operator-scripts\") pod \"placement-b756-account-create-n4p6g\" (UID: \"ff776fa9-dc41-4535-8a27-579d82db15b1\") " pod="openstack/placement-b756-account-create-n4p6g" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.386521 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqdvk\" (UniqueName: \"kubernetes.io/projected/501b7faf-95bc-41d8-8b80-20c1c144395a-kube-api-access-bqdvk\") pod \"placement-db-create-wg496\" (UID: \"501b7faf-95bc-41d8-8b80-20c1c144395a\") " pod="openstack/placement-db-create-wg496" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.386601 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/501b7faf-95bc-41d8-8b80-20c1c144395a-operator-scripts\") pod \"placement-db-create-wg496\" (UID: \"501b7faf-95bc-41d8-8b80-20c1c144395a\") " pod="openstack/placement-db-create-wg496" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.386662 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntjss\" (UniqueName: \"kubernetes.io/projected/ff776fa9-dc41-4535-8a27-579d82db15b1-kube-api-access-ntjss\") pod \"placement-b756-account-create-n4p6g\" (UID: \"ff776fa9-dc41-4535-8a27-579d82db15b1\") " pod="openstack/placement-b756-account-create-n4p6g" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.387576 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/501b7faf-95bc-41d8-8b80-20c1c144395a-operator-scripts\") pod \"placement-db-create-wg496\" (UID: \"501b7faf-95bc-41d8-8b80-20c1c144395a\") " pod="openstack/placement-db-create-wg496" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.405412 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqdvk\" (UniqueName: \"kubernetes.io/projected/501b7faf-95bc-41d8-8b80-20c1c144395a-kube-api-access-bqdvk\") pod \"placement-db-create-wg496\" (UID: \"501b7faf-95bc-41d8-8b80-20c1c144395a\") " pod="openstack/placement-db-create-wg496" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.468439 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.488985 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff776fa9-dc41-4535-8a27-579d82db15b1-operator-scripts\") pod \"placement-b756-account-create-n4p6g\" (UID: \"ff776fa9-dc41-4535-8a27-579d82db15b1\") " pod="openstack/placement-b756-account-create-n4p6g" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.489228 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntjss\" (UniqueName: \"kubernetes.io/projected/ff776fa9-dc41-4535-8a27-579d82db15b1-kube-api-access-ntjss\") pod \"placement-b756-account-create-n4p6g\" (UID: \"ff776fa9-dc41-4535-8a27-579d82db15b1\") " pod="openstack/placement-b756-account-create-n4p6g" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.490140 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff776fa9-dc41-4535-8a27-579d82db15b1-operator-scripts\") pod \"placement-b756-account-create-n4p6g\" (UID: \"ff776fa9-dc41-4535-8a27-579d82db15b1\") " pod="openstack/placement-b756-account-create-n4p6g" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.514560 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-jbncz"] Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.515842 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jbncz" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.534533 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntjss\" (UniqueName: \"kubernetes.io/projected/ff776fa9-dc41-4535-8a27-579d82db15b1-kube-api-access-ntjss\") pod \"placement-b756-account-create-n4p6g\" (UID: \"ff776fa9-dc41-4535-8a27-579d82db15b1\") " pod="openstack/placement-b756-account-create-n4p6g" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.559884 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-jbncz"] Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.588215 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-4zbcq"] Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.590903 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6zcs\" (UniqueName: \"kubernetes.io/projected/64ae0a6e-1bb1-4db6-b021-b46253c6bbb0-kube-api-access-r6zcs\") pod \"glance-db-create-jbncz\" (UID: \"64ae0a6e-1bb1-4db6-b021-b46253c6bbb0\") " pod="openstack/glance-db-create-jbncz" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.590953 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64ae0a6e-1bb1-4db6-b021-b46253c6bbb0-operator-scripts\") pod \"glance-db-create-jbncz\" (UID: \"64ae0a6e-1bb1-4db6-b021-b46253c6bbb0\") " pod="openstack/glance-db-create-jbncz" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.607046 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wg496" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.664081 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-0c45-account-create-bcpj2"] Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.665148 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0c45-account-create-bcpj2" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.667852 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.669882 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-0c45-account-create-bcpj2"] Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.692015 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99tkm\" (UniqueName: \"kubernetes.io/projected/1a30336b-d279-4487-9748-e32b0332282b-kube-api-access-99tkm\") pod \"glance-0c45-account-create-bcpj2\" (UID: \"1a30336b-d279-4487-9748-e32b0332282b\") " pod="openstack/glance-0c45-account-create-bcpj2" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.692076 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6zcs\" (UniqueName: \"kubernetes.io/projected/64ae0a6e-1bb1-4db6-b021-b46253c6bbb0-kube-api-access-r6zcs\") pod \"glance-db-create-jbncz\" (UID: \"64ae0a6e-1bb1-4db6-b021-b46253c6bbb0\") " pod="openstack/glance-db-create-jbncz" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.692108 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64ae0a6e-1bb1-4db6-b021-b46253c6bbb0-operator-scripts\") pod \"glance-db-create-jbncz\" (UID: \"64ae0a6e-1bb1-4db6-b021-b46253c6bbb0\") " pod="openstack/glance-db-create-jbncz" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.692144 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a30336b-d279-4487-9748-e32b0332282b-operator-scripts\") pod \"glance-0c45-account-create-bcpj2\" (UID: \"1a30336b-d279-4487-9748-e32b0332282b\") " pod="openstack/glance-0c45-account-create-bcpj2" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.698240 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64ae0a6e-1bb1-4db6-b021-b46253c6bbb0-operator-scripts\") pod \"glance-db-create-jbncz\" (UID: \"64ae0a6e-1bb1-4db6-b021-b46253c6bbb0\") " pod="openstack/glance-db-create-jbncz" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.715675 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6zcs\" (UniqueName: \"kubernetes.io/projected/64ae0a6e-1bb1-4db6-b021-b46253c6bbb0-kube-api-access-r6zcs\") pod \"glance-db-create-jbncz\" (UID: \"64ae0a6e-1bb1-4db6-b021-b46253c6bbb0\") " pod="openstack/glance-db-create-jbncz" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.794050 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99tkm\" (UniqueName: \"kubernetes.io/projected/1a30336b-d279-4487-9748-e32b0332282b-kube-api-access-99tkm\") pod \"glance-0c45-account-create-bcpj2\" (UID: \"1a30336b-d279-4487-9748-e32b0332282b\") " pod="openstack/glance-0c45-account-create-bcpj2" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.794137 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a30336b-d279-4487-9748-e32b0332282b-operator-scripts\") pod \"glance-0c45-account-create-bcpj2\" (UID: \"1a30336b-d279-4487-9748-e32b0332282b\") " pod="openstack/glance-0c45-account-create-bcpj2" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.802896 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b756-account-create-n4p6g" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.803454 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a30336b-d279-4487-9748-e32b0332282b-operator-scripts\") pod \"glance-0c45-account-create-bcpj2\" (UID: \"1a30336b-d279-4487-9748-e32b0332282b\") " pod="openstack/glance-0c45-account-create-bcpj2" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.817023 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99tkm\" (UniqueName: \"kubernetes.io/projected/1a30336b-d279-4487-9748-e32b0332282b-kube-api-access-99tkm\") pod \"glance-0c45-account-create-bcpj2\" (UID: \"1a30336b-d279-4487-9748-e32b0332282b\") " pod="openstack/glance-0c45-account-create-bcpj2" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.857188 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jbncz" Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.866390 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" podUID="2a26353d-d3de-497e-a6b0-260ca361ca1d" containerName="dnsmasq-dns" containerID="cri-o://5d4a3c160d8959308cc0165218bb8933482204d3fa3f9e867d62f3fe3f93fc56" gracePeriod=10 Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.866593 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-4fc8-account-create-4nkhn"] Nov 22 05:09:50 crc kubenswrapper[4818]: W1122 05:09:50.886666 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b0bbbc5_6ad1_4390_8d2b_b91c83f49701.slice/crio-f65625bef84634fdefa7ff09a93369d9f215a3126bad4718137723ec1d62bb7d WatchSource:0}: Error finding container f65625bef84634fdefa7ff09a93369d9f215a3126bad4718137723ec1d62bb7d: Status 404 returned error can't find the container with id f65625bef84634fdefa7ff09a93369d9f215a3126bad4718137723ec1d62bb7d Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.889815 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-wg496"] Nov 22 05:09:50 crc kubenswrapper[4818]: W1122 05:09:50.908136 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod501b7faf_95bc_41d8_8b80_20c1c144395a.slice/crio-5d5c616cb1749104ca0ffb90fe1a94975d4bf4ac2d7fcc7252146db8583d08ad WatchSource:0}: Error finding container 5d5c616cb1749104ca0ffb90fe1a94975d4bf4ac2d7fcc7252146db8583d08ad: Status 404 returned error can't find the container with id 5d5c616cb1749104ca0ffb90fe1a94975d4bf4ac2d7fcc7252146db8583d08ad Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.957833 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gmlh9"] Nov 22 05:09:50 crc kubenswrapper[4818]: W1122 05:09:50.985087 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f234b24_4e7a_42dc_943d_aa45cb9cd72b.slice/crio-8c2a8cdc421d508aeb3638b063b50acb219290e3e8284baa4ca3061b7c1fce64 WatchSource:0}: Error finding container 8c2a8cdc421d508aeb3638b063b50acb219290e3e8284baa4ca3061b7c1fce64: Status 404 returned error can't find the container with id 8c2a8cdc421d508aeb3638b063b50acb219290e3e8284baa4ca3061b7c1fce64 Nov 22 05:09:50 crc kubenswrapper[4818]: I1122 05:09:50.985153 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0c45-account-create-bcpj2" Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.298790 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-b756-account-create-n4p6g"] Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.379175 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-jbncz"] Nov 22 05:09:51 crc kubenswrapper[4818]: W1122 05:09:51.389877 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64ae0a6e_1bb1_4db6_b021_b46253c6bbb0.slice/crio-b59b2b263e83535c5a6bf7a839cf20ccf99995c707930a20ee89b07e2f605a20 WatchSource:0}: Error finding container b59b2b263e83535c5a6bf7a839cf20ccf99995c707930a20ee89b07e2f605a20: Status 404 returned error can't find the container with id b59b2b263e83535c5a6bf7a839cf20ccf99995c707930a20ee89b07e2f605a20 Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.487825 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-0c45-account-create-bcpj2"] Nov 22 05:09:51 crc kubenswrapper[4818]: W1122 05:09:51.493108 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a30336b_d279_4487_9748_e32b0332282b.slice/crio-442381c5e82895f2d3d4c4a25a18a19a9126fa287371fac7081262ca9df6b707 WatchSource:0}: Error finding container 442381c5e82895f2d3d4c4a25a18a19a9126fa287371fac7081262ca9df6b707: Status 404 returned error can't find the container with id 442381c5e82895f2d3d4c4a25a18a19a9126fa287371fac7081262ca9df6b707 Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.746780 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.811142 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a26353d-d3de-497e-a6b0-260ca361ca1d-config\") pod \"2a26353d-d3de-497e-a6b0-260ca361ca1d\" (UID: \"2a26353d-d3de-497e-a6b0-260ca361ca1d\") " Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.811274 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a26353d-d3de-497e-a6b0-260ca361ca1d-ovsdbserver-nb\") pod \"2a26353d-d3de-497e-a6b0-260ca361ca1d\" (UID: \"2a26353d-d3de-497e-a6b0-260ca361ca1d\") " Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.811348 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lg8k9\" (UniqueName: \"kubernetes.io/projected/2a26353d-d3de-497e-a6b0-260ca361ca1d-kube-api-access-lg8k9\") pod \"2a26353d-d3de-497e-a6b0-260ca361ca1d\" (UID: \"2a26353d-d3de-497e-a6b0-260ca361ca1d\") " Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.811428 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a26353d-d3de-497e-a6b0-260ca361ca1d-dns-svc\") pod \"2a26353d-d3de-497e-a6b0-260ca361ca1d\" (UID: \"2a26353d-d3de-497e-a6b0-260ca361ca1d\") " Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.817650 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a26353d-d3de-497e-a6b0-260ca361ca1d-kube-api-access-lg8k9" (OuterVolumeSpecName: "kube-api-access-lg8k9") pod "2a26353d-d3de-497e-a6b0-260ca361ca1d" (UID: "2a26353d-d3de-497e-a6b0-260ca361ca1d"). InnerVolumeSpecName "kube-api-access-lg8k9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.872068 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a26353d-d3de-497e-a6b0-260ca361ca1d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2a26353d-d3de-497e-a6b0-260ca361ca1d" (UID: "2a26353d-d3de-497e-a6b0-260ca361ca1d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.874667 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a26353d-d3de-497e-a6b0-260ca361ca1d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2a26353d-d3de-497e-a6b0-260ca361ca1d" (UID: "2a26353d-d3de-497e-a6b0-260ca361ca1d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.879985 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0c45-account-create-bcpj2" event={"ID":"1a30336b-d279-4487-9748-e32b0332282b","Type":"ContainerStarted","Data":"26aec8102cb98ea8e36f0a2ee42ecc1ca4b15ea1499c6e4b0f381b676f0997da"} Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.880032 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0c45-account-create-bcpj2" event={"ID":"1a30336b-d279-4487-9748-e32b0332282b","Type":"ContainerStarted","Data":"442381c5e82895f2d3d4c4a25a18a19a9126fa287371fac7081262ca9df6b707"} Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.881337 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a26353d-d3de-497e-a6b0-260ca361ca1d-config" (OuterVolumeSpecName: "config") pod "2a26353d-d3de-497e-a6b0-260ca361ca1d" (UID: "2a26353d-d3de-497e-a6b0-260ca361ca1d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.896095 4818 generic.go:334] "Generic (PLEG): container finished" podID="5b0bbbc5-6ad1-4390-8d2b-b91c83f49701" containerID="e90262f97d27864977d3acc2ab0a4b913d5d787ea950db716fedd7f07af83b64" exitCode=0 Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.896193 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4fc8-account-create-4nkhn" event={"ID":"5b0bbbc5-6ad1-4390-8d2b-b91c83f49701","Type":"ContainerDied","Data":"e90262f97d27864977d3acc2ab0a4b913d5d787ea950db716fedd7f07af83b64"} Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.896242 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4fc8-account-create-4nkhn" event={"ID":"5b0bbbc5-6ad1-4390-8d2b-b91c83f49701","Type":"ContainerStarted","Data":"f65625bef84634fdefa7ff09a93369d9f215a3126bad4718137723ec1d62bb7d"} Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.896825 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-0c45-account-create-bcpj2" podStartSLOduration=1.8968023619999999 podStartE2EDuration="1.896802362s" podCreationTimestamp="2025-11-22 05:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:09:51.895152239 +0000 UTC m=+1344.469568776" watchObservedRunningTime="2025-11-22 05:09:51.896802362 +0000 UTC m=+1344.471218889" Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.899637 4818 generic.go:334] "Generic (PLEG): container finished" podID="501b7faf-95bc-41d8-8b80-20c1c144395a" containerID="294ae3c59b977c4294755965dc0a6cf3c4693284f0a330bd64b4f1cc10c48cc2" exitCode=0 Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.899706 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wg496" event={"ID":"501b7faf-95bc-41d8-8b80-20c1c144395a","Type":"ContainerDied","Data":"294ae3c59b977c4294755965dc0a6cf3c4693284f0a330bd64b4f1cc10c48cc2"} Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.899735 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wg496" event={"ID":"501b7faf-95bc-41d8-8b80-20c1c144395a","Type":"ContainerStarted","Data":"5d5c616cb1749104ca0ffb90fe1a94975d4bf4ac2d7fcc7252146db8583d08ad"} Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.901285 4818 generic.go:334] "Generic (PLEG): container finished" podID="2a26353d-d3de-497e-a6b0-260ca361ca1d" containerID="5d4a3c160d8959308cc0165218bb8933482204d3fa3f9e867d62f3fe3f93fc56" exitCode=0 Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.901335 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" event={"ID":"2a26353d-d3de-497e-a6b0-260ca361ca1d","Type":"ContainerDied","Data":"5d4a3c160d8959308cc0165218bb8933482204d3fa3f9e867d62f3fe3f93fc56"} Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.901357 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" event={"ID":"2a26353d-d3de-497e-a6b0-260ca361ca1d","Type":"ContainerDied","Data":"cc64b8500b3d94f041e4830dc12232b54bbcf022cb5531ae70bff2e98af4fc11"} Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.901377 4818 scope.go:117] "RemoveContainer" containerID="5d4a3c160d8959308cc0165218bb8933482204d3fa3f9e867d62f3fe3f93fc56" Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.901506 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-4zbcq" Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.905721 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-jbncz" event={"ID":"64ae0a6e-1bb1-4db6-b021-b46253c6bbb0","Type":"ContainerStarted","Data":"07f7237b64c16e30208f0f923aae7adde73e12131e1c38130a1593ff036fd433"} Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.905752 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-jbncz" event={"ID":"64ae0a6e-1bb1-4db6-b021-b46253c6bbb0","Type":"ContainerStarted","Data":"b59b2b263e83535c5a6bf7a839cf20ccf99995c707930a20ee89b07e2f605a20"} Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.912417 4818 generic.go:334] "Generic (PLEG): container finished" podID="4f234b24-4e7a-42dc-943d-aa45cb9cd72b" containerID="983d3e8a99dd527229e00713cbe39c9cfe0a090d0d47ecd5f7c7ee9e23d16dd3" exitCode=0 Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.912451 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gmlh9" event={"ID":"4f234b24-4e7a-42dc-943d-aa45cb9cd72b","Type":"ContainerDied","Data":"983d3e8a99dd527229e00713cbe39c9cfe0a090d0d47ecd5f7c7ee9e23d16dd3"} Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.912707 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gmlh9" event={"ID":"4f234b24-4e7a-42dc-943d-aa45cb9cd72b","Type":"ContainerStarted","Data":"8c2a8cdc421d508aeb3638b063b50acb219290e3e8284baa4ca3061b7c1fce64"} Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.913640 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a26353d-d3de-497e-a6b0-260ca361ca1d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.914190 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lg8k9\" (UniqueName: \"kubernetes.io/projected/2a26353d-d3de-497e-a6b0-260ca361ca1d-kube-api-access-lg8k9\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.914206 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a26353d-d3de-497e-a6b0-260ca361ca1d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.914219 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a26353d-d3de-497e-a6b0-260ca361ca1d-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.917436 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b756-account-create-n4p6g" event={"ID":"ff776fa9-dc41-4535-8a27-579d82db15b1","Type":"ContainerStarted","Data":"f338a44aab1fde1e2024f62e698dea2702f40d45d9247f882f874f7f4f861e98"} Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.917479 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b756-account-create-n4p6g" event={"ID":"ff776fa9-dc41-4535-8a27-579d82db15b1","Type":"ContainerStarted","Data":"0d3f1f98bd393af0196c387e762b06ba3efc2e3501076aa6277bc33660d68117"} Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.926348 4818 scope.go:117] "RemoveContainer" containerID="878ab50cf31c6349bdb71aea880bc54b3848abe72d9af7ad74c6e73510ab7788" Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.977622 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-jbncz" podStartSLOduration=1.97760361 podStartE2EDuration="1.97760361s" podCreationTimestamp="2025-11-22 05:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:09:51.963093314 +0000 UTC m=+1344.537509841" watchObservedRunningTime="2025-11-22 05:09:51.97760361 +0000 UTC m=+1344.552020137" Nov 22 05:09:51 crc kubenswrapper[4818]: I1122 05:09:51.978617 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-b756-account-create-n4p6g" podStartSLOduration=1.978609476 podStartE2EDuration="1.978609476s" podCreationTimestamp="2025-11-22 05:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:09:51.9768688 +0000 UTC m=+1344.551285337" watchObservedRunningTime="2025-11-22 05:09:51.978609476 +0000 UTC m=+1344.553026003" Nov 22 05:09:52 crc kubenswrapper[4818]: I1122 05:09:52.007904 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-4zbcq"] Nov 22 05:09:52 crc kubenswrapper[4818]: I1122 05:09:52.013622 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-4zbcq"] Nov 22 05:09:52 crc kubenswrapper[4818]: I1122 05:09:52.118404 4818 scope.go:117] "RemoveContainer" containerID="5d4a3c160d8959308cc0165218bb8933482204d3fa3f9e867d62f3fe3f93fc56" Nov 22 05:09:52 crc kubenswrapper[4818]: E1122 05:09:52.118912 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d4a3c160d8959308cc0165218bb8933482204d3fa3f9e867d62f3fe3f93fc56\": container with ID starting with 5d4a3c160d8959308cc0165218bb8933482204d3fa3f9e867d62f3fe3f93fc56 not found: ID does not exist" containerID="5d4a3c160d8959308cc0165218bb8933482204d3fa3f9e867d62f3fe3f93fc56" Nov 22 05:09:52 crc kubenswrapper[4818]: I1122 05:09:52.118957 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d4a3c160d8959308cc0165218bb8933482204d3fa3f9e867d62f3fe3f93fc56"} err="failed to get container status \"5d4a3c160d8959308cc0165218bb8933482204d3fa3f9e867d62f3fe3f93fc56\": rpc error: code = NotFound desc = could not find container \"5d4a3c160d8959308cc0165218bb8933482204d3fa3f9e867d62f3fe3f93fc56\": container with ID starting with 5d4a3c160d8959308cc0165218bb8933482204d3fa3f9e867d62f3fe3f93fc56 not found: ID does not exist" Nov 22 05:09:52 crc kubenswrapper[4818]: I1122 05:09:52.118984 4818 scope.go:117] "RemoveContainer" containerID="878ab50cf31c6349bdb71aea880bc54b3848abe72d9af7ad74c6e73510ab7788" Nov 22 05:09:52 crc kubenswrapper[4818]: E1122 05:09:52.119764 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"878ab50cf31c6349bdb71aea880bc54b3848abe72d9af7ad74c6e73510ab7788\": container with ID starting with 878ab50cf31c6349bdb71aea880bc54b3848abe72d9af7ad74c6e73510ab7788 not found: ID does not exist" containerID="878ab50cf31c6349bdb71aea880bc54b3848abe72d9af7ad74c6e73510ab7788" Nov 22 05:09:52 crc kubenswrapper[4818]: I1122 05:09:52.119798 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"878ab50cf31c6349bdb71aea880bc54b3848abe72d9af7ad74c6e73510ab7788"} err="failed to get container status \"878ab50cf31c6349bdb71aea880bc54b3848abe72d9af7ad74c6e73510ab7788\": rpc error: code = NotFound desc = could not find container \"878ab50cf31c6349bdb71aea880bc54b3848abe72d9af7ad74c6e73510ab7788\": container with ID starting with 878ab50cf31c6349bdb71aea880bc54b3848abe72d9af7ad74c6e73510ab7788 not found: ID does not exist" Nov 22 05:09:52 crc kubenswrapper[4818]: I1122 05:09:52.309832 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a26353d-d3de-497e-a6b0-260ca361ca1d" path="/var/lib/kubelet/pods/2a26353d-d3de-497e-a6b0-260ca361ca1d/volumes" Nov 22 05:09:52 crc kubenswrapper[4818]: I1122 05:09:52.926445 4818 generic.go:334] "Generic (PLEG): container finished" podID="ff776fa9-dc41-4535-8a27-579d82db15b1" containerID="f338a44aab1fde1e2024f62e698dea2702f40d45d9247f882f874f7f4f861e98" exitCode=0 Nov 22 05:09:52 crc kubenswrapper[4818]: I1122 05:09:52.926503 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b756-account-create-n4p6g" event={"ID":"ff776fa9-dc41-4535-8a27-579d82db15b1","Type":"ContainerDied","Data":"f338a44aab1fde1e2024f62e698dea2702f40d45d9247f882f874f7f4f861e98"} Nov 22 05:09:52 crc kubenswrapper[4818]: I1122 05:09:52.928679 4818 generic.go:334] "Generic (PLEG): container finished" podID="1a30336b-d279-4487-9748-e32b0332282b" containerID="26aec8102cb98ea8e36f0a2ee42ecc1ca4b15ea1499c6e4b0f381b676f0997da" exitCode=0 Nov 22 05:09:52 crc kubenswrapper[4818]: I1122 05:09:52.928729 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0c45-account-create-bcpj2" event={"ID":"1a30336b-d279-4487-9748-e32b0332282b","Type":"ContainerDied","Data":"26aec8102cb98ea8e36f0a2ee42ecc1ca4b15ea1499c6e4b0f381b676f0997da"} Nov 22 05:09:52 crc kubenswrapper[4818]: I1122 05:09:52.932228 4818 generic.go:334] "Generic (PLEG): container finished" podID="64ae0a6e-1bb1-4db6-b021-b46253c6bbb0" containerID="07f7237b64c16e30208f0f923aae7adde73e12131e1c38130a1593ff036fd433" exitCode=0 Nov 22 05:09:52 crc kubenswrapper[4818]: I1122 05:09:52.932416 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-jbncz" event={"ID":"64ae0a6e-1bb1-4db6-b021-b46253c6bbb0","Type":"ContainerDied","Data":"07f7237b64c16e30208f0f923aae7adde73e12131e1c38130a1593ff036fd433"} Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.393747 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4fc8-account-create-4nkhn" Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.401790 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gmlh9" Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.412021 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wg496" Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.540704 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9hxz\" (UniqueName: \"kubernetes.io/projected/5b0bbbc5-6ad1-4390-8d2b-b91c83f49701-kube-api-access-w9hxz\") pod \"5b0bbbc5-6ad1-4390-8d2b-b91c83f49701\" (UID: \"5b0bbbc5-6ad1-4390-8d2b-b91c83f49701\") " Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.540772 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b0bbbc5-6ad1-4390-8d2b-b91c83f49701-operator-scripts\") pod \"5b0bbbc5-6ad1-4390-8d2b-b91c83f49701\" (UID: \"5b0bbbc5-6ad1-4390-8d2b-b91c83f49701\") " Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.540921 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dt9pd\" (UniqueName: \"kubernetes.io/projected/4f234b24-4e7a-42dc-943d-aa45cb9cd72b-kube-api-access-dt9pd\") pod \"4f234b24-4e7a-42dc-943d-aa45cb9cd72b\" (UID: \"4f234b24-4e7a-42dc-943d-aa45cb9cd72b\") " Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.540947 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqdvk\" (UniqueName: \"kubernetes.io/projected/501b7faf-95bc-41d8-8b80-20c1c144395a-kube-api-access-bqdvk\") pod \"501b7faf-95bc-41d8-8b80-20c1c144395a\" (UID: \"501b7faf-95bc-41d8-8b80-20c1c144395a\") " Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.541043 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f234b24-4e7a-42dc-943d-aa45cb9cd72b-operator-scripts\") pod \"4f234b24-4e7a-42dc-943d-aa45cb9cd72b\" (UID: \"4f234b24-4e7a-42dc-943d-aa45cb9cd72b\") " Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.541063 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/501b7faf-95bc-41d8-8b80-20c1c144395a-operator-scripts\") pod \"501b7faf-95bc-41d8-8b80-20c1c144395a\" (UID: \"501b7faf-95bc-41d8-8b80-20c1c144395a\") " Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.542189 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f234b24-4e7a-42dc-943d-aa45cb9cd72b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4f234b24-4e7a-42dc-943d-aa45cb9cd72b" (UID: "4f234b24-4e7a-42dc-943d-aa45cb9cd72b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.542292 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/501b7faf-95bc-41d8-8b80-20c1c144395a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "501b7faf-95bc-41d8-8b80-20c1c144395a" (UID: "501b7faf-95bc-41d8-8b80-20c1c144395a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.542336 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b0bbbc5-6ad1-4390-8d2b-b91c83f49701-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5b0bbbc5-6ad1-4390-8d2b-b91c83f49701" (UID: "5b0bbbc5-6ad1-4390-8d2b-b91c83f49701"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.545054 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f234b24-4e7a-42dc-943d-aa45cb9cd72b-kube-api-access-dt9pd" (OuterVolumeSpecName: "kube-api-access-dt9pd") pod "4f234b24-4e7a-42dc-943d-aa45cb9cd72b" (UID: "4f234b24-4e7a-42dc-943d-aa45cb9cd72b"). InnerVolumeSpecName "kube-api-access-dt9pd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.545100 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/501b7faf-95bc-41d8-8b80-20c1c144395a-kube-api-access-bqdvk" (OuterVolumeSpecName: "kube-api-access-bqdvk") pod "501b7faf-95bc-41d8-8b80-20c1c144395a" (UID: "501b7faf-95bc-41d8-8b80-20c1c144395a"). InnerVolumeSpecName "kube-api-access-bqdvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.549205 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b0bbbc5-6ad1-4390-8d2b-b91c83f49701-kube-api-access-w9hxz" (OuterVolumeSpecName: "kube-api-access-w9hxz") pod "5b0bbbc5-6ad1-4390-8d2b-b91c83f49701" (UID: "5b0bbbc5-6ad1-4390-8d2b-b91c83f49701"). InnerVolumeSpecName "kube-api-access-w9hxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.642719 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f234b24-4e7a-42dc-943d-aa45cb9cd72b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.642769 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/501b7faf-95bc-41d8-8b80-20c1c144395a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.642787 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9hxz\" (UniqueName: \"kubernetes.io/projected/5b0bbbc5-6ad1-4390-8d2b-b91c83f49701-kube-api-access-w9hxz\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.642808 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b0bbbc5-6ad1-4390-8d2b-b91c83f49701-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.642824 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dt9pd\" (UniqueName: \"kubernetes.io/projected/4f234b24-4e7a-42dc-943d-aa45cb9cd72b-kube-api-access-dt9pd\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.642838 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqdvk\" (UniqueName: \"kubernetes.io/projected/501b7faf-95bc-41d8-8b80-20c1c144395a-kube-api-access-bqdvk\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.944851 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gmlh9" event={"ID":"4f234b24-4e7a-42dc-943d-aa45cb9cd72b","Type":"ContainerDied","Data":"8c2a8cdc421d508aeb3638b063b50acb219290e3e8284baa4ca3061b7c1fce64"} Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.945201 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c2a8cdc421d508aeb3638b063b50acb219290e3e8284baa4ca3061b7c1fce64" Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.945001 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gmlh9" Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.947462 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4fc8-account-create-4nkhn" event={"ID":"5b0bbbc5-6ad1-4390-8d2b-b91c83f49701","Type":"ContainerDied","Data":"f65625bef84634fdefa7ff09a93369d9f215a3126bad4718137723ec1d62bb7d"} Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.947515 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f65625bef84634fdefa7ff09a93369d9f215a3126bad4718137723ec1d62bb7d" Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.947488 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4fc8-account-create-4nkhn" Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.949538 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wg496" event={"ID":"501b7faf-95bc-41d8-8b80-20c1c144395a","Type":"ContainerDied","Data":"5d5c616cb1749104ca0ffb90fe1a94975d4bf4ac2d7fcc7252146db8583d08ad"} Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.949577 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d5c616cb1749104ca0ffb90fe1a94975d4bf4ac2d7fcc7252146db8583d08ad" Nov 22 05:09:53 crc kubenswrapper[4818]: I1122 05:09:53.949741 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wg496" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.217011 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b756-account-create-n4p6g" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.314025 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jbncz" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.353463 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntjss\" (UniqueName: \"kubernetes.io/projected/ff776fa9-dc41-4535-8a27-579d82db15b1-kube-api-access-ntjss\") pod \"ff776fa9-dc41-4535-8a27-579d82db15b1\" (UID: \"ff776fa9-dc41-4535-8a27-579d82db15b1\") " Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.353700 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff776fa9-dc41-4535-8a27-579d82db15b1-operator-scripts\") pod \"ff776fa9-dc41-4535-8a27-579d82db15b1\" (UID: \"ff776fa9-dc41-4535-8a27-579d82db15b1\") " Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.354600 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff776fa9-dc41-4535-8a27-579d82db15b1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ff776fa9-dc41-4535-8a27-579d82db15b1" (UID: "ff776fa9-dc41-4535-8a27-579d82db15b1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.359049 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff776fa9-dc41-4535-8a27-579d82db15b1-kube-api-access-ntjss" (OuterVolumeSpecName: "kube-api-access-ntjss") pod "ff776fa9-dc41-4535-8a27-579d82db15b1" (UID: "ff776fa9-dc41-4535-8a27-579d82db15b1"). InnerVolumeSpecName "kube-api-access-ntjss". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.446565 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0c45-account-create-bcpj2" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.455667 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64ae0a6e-1bb1-4db6-b021-b46253c6bbb0-operator-scripts\") pod \"64ae0a6e-1bb1-4db6-b021-b46253c6bbb0\" (UID: \"64ae0a6e-1bb1-4db6-b021-b46253c6bbb0\") " Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.455753 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6zcs\" (UniqueName: \"kubernetes.io/projected/64ae0a6e-1bb1-4db6-b021-b46253c6bbb0-kube-api-access-r6zcs\") pod \"64ae0a6e-1bb1-4db6-b021-b46253c6bbb0\" (UID: \"64ae0a6e-1bb1-4db6-b021-b46253c6bbb0\") " Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.456184 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64ae0a6e-1bb1-4db6-b021-b46253c6bbb0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "64ae0a6e-1bb1-4db6-b021-b46253c6bbb0" (UID: "64ae0a6e-1bb1-4db6-b021-b46253c6bbb0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.456224 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff776fa9-dc41-4535-8a27-579d82db15b1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.456246 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntjss\" (UniqueName: \"kubernetes.io/projected/ff776fa9-dc41-4535-8a27-579d82db15b1-kube-api-access-ntjss\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.458542 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64ae0a6e-1bb1-4db6-b021-b46253c6bbb0-kube-api-access-r6zcs" (OuterVolumeSpecName: "kube-api-access-r6zcs") pod "64ae0a6e-1bb1-4db6-b021-b46253c6bbb0" (UID: "64ae0a6e-1bb1-4db6-b021-b46253c6bbb0"). InnerVolumeSpecName "kube-api-access-r6zcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.557682 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99tkm\" (UniqueName: \"kubernetes.io/projected/1a30336b-d279-4487-9748-e32b0332282b-kube-api-access-99tkm\") pod \"1a30336b-d279-4487-9748-e32b0332282b\" (UID: \"1a30336b-d279-4487-9748-e32b0332282b\") " Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.557849 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a30336b-d279-4487-9748-e32b0332282b-operator-scripts\") pod \"1a30336b-d279-4487-9748-e32b0332282b\" (UID: \"1a30336b-d279-4487-9748-e32b0332282b\") " Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.558195 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64ae0a6e-1bb1-4db6-b021-b46253c6bbb0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.558215 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6zcs\" (UniqueName: \"kubernetes.io/projected/64ae0a6e-1bb1-4db6-b021-b46253c6bbb0-kube-api-access-r6zcs\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.558402 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a30336b-d279-4487-9748-e32b0332282b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1a30336b-d279-4487-9748-e32b0332282b" (UID: "1a30336b-d279-4487-9748-e32b0332282b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.561874 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a30336b-d279-4487-9748-e32b0332282b-kube-api-access-99tkm" (OuterVolumeSpecName: "kube-api-access-99tkm") pod "1a30336b-d279-4487-9748-e32b0332282b" (UID: "1a30336b-d279-4487-9748-e32b0332282b"). InnerVolumeSpecName "kube-api-access-99tkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.659882 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a30336b-d279-4487-9748-e32b0332282b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.659912 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99tkm\" (UniqueName: \"kubernetes.io/projected/1a30336b-d279-4487-9748-e32b0332282b-kube-api-access-99tkm\") on node \"crc\" DevicePath \"\"" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.962824 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b756-account-create-n4p6g" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.962844 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b756-account-create-n4p6g" event={"ID":"ff776fa9-dc41-4535-8a27-579d82db15b1","Type":"ContainerDied","Data":"0d3f1f98bd393af0196c387e762b06ba3efc2e3501076aa6277bc33660d68117"} Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.962936 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d3f1f98bd393af0196c387e762b06ba3efc2e3501076aa6277bc33660d68117" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.965763 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0c45-account-create-bcpj2" event={"ID":"1a30336b-d279-4487-9748-e32b0332282b","Type":"ContainerDied","Data":"442381c5e82895f2d3d4c4a25a18a19a9126fa287371fac7081262ca9df6b707"} Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.965827 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="442381c5e82895f2d3d4c4a25a18a19a9126fa287371fac7081262ca9df6b707" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.965776 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0c45-account-create-bcpj2" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.968345 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-jbncz" event={"ID":"64ae0a6e-1bb1-4db6-b021-b46253c6bbb0","Type":"ContainerDied","Data":"b59b2b263e83535c5a6bf7a839cf20ccf99995c707930a20ee89b07e2f605a20"} Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.968393 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b59b2b263e83535c5a6bf7a839cf20ccf99995c707930a20ee89b07e2f605a20" Nov 22 05:09:54 crc kubenswrapper[4818]: I1122 05:09:54.968627 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jbncz" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.414905 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.916195 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-jpkvr"] Nov 22 05:09:55 crc kubenswrapper[4818]: E1122 05:09:55.916509 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501b7faf-95bc-41d8-8b80-20c1c144395a" containerName="mariadb-database-create" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.916521 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="501b7faf-95bc-41d8-8b80-20c1c144395a" containerName="mariadb-database-create" Nov 22 05:09:55 crc kubenswrapper[4818]: E1122 05:09:55.916538 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64ae0a6e-1bb1-4db6-b021-b46253c6bbb0" containerName="mariadb-database-create" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.916543 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="64ae0a6e-1bb1-4db6-b021-b46253c6bbb0" containerName="mariadb-database-create" Nov 22 05:09:55 crc kubenswrapper[4818]: E1122 05:09:55.916552 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a26353d-d3de-497e-a6b0-260ca361ca1d" containerName="dnsmasq-dns" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.916559 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a26353d-d3de-497e-a6b0-260ca361ca1d" containerName="dnsmasq-dns" Nov 22 05:09:55 crc kubenswrapper[4818]: E1122 05:09:55.916576 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff776fa9-dc41-4535-8a27-579d82db15b1" containerName="mariadb-account-create" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.916581 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff776fa9-dc41-4535-8a27-579d82db15b1" containerName="mariadb-account-create" Nov 22 05:09:55 crc kubenswrapper[4818]: E1122 05:09:55.916590 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f234b24-4e7a-42dc-943d-aa45cb9cd72b" containerName="mariadb-database-create" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.916596 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f234b24-4e7a-42dc-943d-aa45cb9cd72b" containerName="mariadb-database-create" Nov 22 05:09:55 crc kubenswrapper[4818]: E1122 05:09:55.916607 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a30336b-d279-4487-9748-e32b0332282b" containerName="mariadb-account-create" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.916614 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a30336b-d279-4487-9748-e32b0332282b" containerName="mariadb-account-create" Nov 22 05:09:55 crc kubenswrapper[4818]: E1122 05:09:55.916623 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b0bbbc5-6ad1-4390-8d2b-b91c83f49701" containerName="mariadb-account-create" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.916629 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b0bbbc5-6ad1-4390-8d2b-b91c83f49701" containerName="mariadb-account-create" Nov 22 05:09:55 crc kubenswrapper[4818]: E1122 05:09:55.916638 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a26353d-d3de-497e-a6b0-260ca361ca1d" containerName="init" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.916644 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a26353d-d3de-497e-a6b0-260ca361ca1d" containerName="init" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.916779 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b0bbbc5-6ad1-4390-8d2b-b91c83f49701" containerName="mariadb-account-create" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.916801 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="501b7faf-95bc-41d8-8b80-20c1c144395a" containerName="mariadb-database-create" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.916816 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a30336b-d279-4487-9748-e32b0332282b" containerName="mariadb-account-create" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.916835 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a26353d-d3de-497e-a6b0-260ca361ca1d" containerName="dnsmasq-dns" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.916846 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f234b24-4e7a-42dc-943d-aa45cb9cd72b" containerName="mariadb-database-create" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.916858 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="64ae0a6e-1bb1-4db6-b021-b46253c6bbb0" containerName="mariadb-database-create" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.916875 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff776fa9-dc41-4535-8a27-579d82db15b1" containerName="mariadb-account-create" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.917364 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jpkvr" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.920422 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2zbds" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.920430 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 22 05:09:55 crc kubenswrapper[4818]: I1122 05:09:55.932589 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-jpkvr"] Nov 22 05:09:56 crc kubenswrapper[4818]: I1122 05:09:56.083321 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/486177c7-b0c3-45ba-a18c-79509154c9df-combined-ca-bundle\") pod \"glance-db-sync-jpkvr\" (UID: \"486177c7-b0c3-45ba-a18c-79509154c9df\") " pod="openstack/glance-db-sync-jpkvr" Nov 22 05:09:56 crc kubenswrapper[4818]: I1122 05:09:56.083501 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/486177c7-b0c3-45ba-a18c-79509154c9df-config-data\") pod \"glance-db-sync-jpkvr\" (UID: \"486177c7-b0c3-45ba-a18c-79509154c9df\") " pod="openstack/glance-db-sync-jpkvr" Nov 22 05:09:56 crc kubenswrapper[4818]: I1122 05:09:56.083865 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/486177c7-b0c3-45ba-a18c-79509154c9df-db-sync-config-data\") pod \"glance-db-sync-jpkvr\" (UID: \"486177c7-b0c3-45ba-a18c-79509154c9df\") " pod="openstack/glance-db-sync-jpkvr" Nov 22 05:09:56 crc kubenswrapper[4818]: I1122 05:09:56.084067 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br247\" (UniqueName: \"kubernetes.io/projected/486177c7-b0c3-45ba-a18c-79509154c9df-kube-api-access-br247\") pod \"glance-db-sync-jpkvr\" (UID: \"486177c7-b0c3-45ba-a18c-79509154c9df\") " pod="openstack/glance-db-sync-jpkvr" Nov 22 05:09:56 crc kubenswrapper[4818]: I1122 05:09:56.185656 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br247\" (UniqueName: \"kubernetes.io/projected/486177c7-b0c3-45ba-a18c-79509154c9df-kube-api-access-br247\") pod \"glance-db-sync-jpkvr\" (UID: \"486177c7-b0c3-45ba-a18c-79509154c9df\") " pod="openstack/glance-db-sync-jpkvr" Nov 22 05:09:56 crc kubenswrapper[4818]: I1122 05:09:56.185828 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/486177c7-b0c3-45ba-a18c-79509154c9df-combined-ca-bundle\") pod \"glance-db-sync-jpkvr\" (UID: \"486177c7-b0c3-45ba-a18c-79509154c9df\") " pod="openstack/glance-db-sync-jpkvr" Nov 22 05:09:56 crc kubenswrapper[4818]: I1122 05:09:56.185901 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/486177c7-b0c3-45ba-a18c-79509154c9df-config-data\") pod \"glance-db-sync-jpkvr\" (UID: \"486177c7-b0c3-45ba-a18c-79509154c9df\") " pod="openstack/glance-db-sync-jpkvr" Nov 22 05:09:56 crc kubenswrapper[4818]: I1122 05:09:56.186030 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/486177c7-b0c3-45ba-a18c-79509154c9df-db-sync-config-data\") pod \"glance-db-sync-jpkvr\" (UID: \"486177c7-b0c3-45ba-a18c-79509154c9df\") " pod="openstack/glance-db-sync-jpkvr" Nov 22 05:09:56 crc kubenswrapper[4818]: I1122 05:09:56.190483 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/486177c7-b0c3-45ba-a18c-79509154c9df-db-sync-config-data\") pod \"glance-db-sync-jpkvr\" (UID: \"486177c7-b0c3-45ba-a18c-79509154c9df\") " pod="openstack/glance-db-sync-jpkvr" Nov 22 05:09:56 crc kubenswrapper[4818]: I1122 05:09:56.191671 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/486177c7-b0c3-45ba-a18c-79509154c9df-combined-ca-bundle\") pod \"glance-db-sync-jpkvr\" (UID: \"486177c7-b0c3-45ba-a18c-79509154c9df\") " pod="openstack/glance-db-sync-jpkvr" Nov 22 05:09:56 crc kubenswrapper[4818]: I1122 05:09:56.192523 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/486177c7-b0c3-45ba-a18c-79509154c9df-config-data\") pod \"glance-db-sync-jpkvr\" (UID: \"486177c7-b0c3-45ba-a18c-79509154c9df\") " pod="openstack/glance-db-sync-jpkvr" Nov 22 05:09:56 crc kubenswrapper[4818]: I1122 05:09:56.208693 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br247\" (UniqueName: \"kubernetes.io/projected/486177c7-b0c3-45ba-a18c-79509154c9df-kube-api-access-br247\") pod \"glance-db-sync-jpkvr\" (UID: \"486177c7-b0c3-45ba-a18c-79509154c9df\") " pod="openstack/glance-db-sync-jpkvr" Nov 22 05:09:56 crc kubenswrapper[4818]: I1122 05:09:56.240303 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jpkvr" Nov 22 05:09:56 crc kubenswrapper[4818]: I1122 05:09:56.884503 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-jpkvr"] Nov 22 05:09:56 crc kubenswrapper[4818]: W1122 05:09:56.887014 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod486177c7_b0c3_45ba_a18c_79509154c9df.slice/crio-41272fc91993fc39171c161d477bebda06b770971e966d3727dbcc6e9d6392a7 WatchSource:0}: Error finding container 41272fc91993fc39171c161d477bebda06b770971e966d3727dbcc6e9d6392a7: Status 404 returned error can't find the container with id 41272fc91993fc39171c161d477bebda06b770971e966d3727dbcc6e9d6392a7 Nov 22 05:09:56 crc kubenswrapper[4818]: I1122 05:09:56.985397 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jpkvr" event={"ID":"486177c7-b0c3-45ba-a18c-79509154c9df","Type":"ContainerStarted","Data":"41272fc91993fc39171c161d477bebda06b770971e966d3727dbcc6e9d6392a7"} Nov 22 05:10:06 crc kubenswrapper[4818]: I1122 05:10:06.078061 4818 generic.go:334] "Generic (PLEG): container finished" podID="69f72526-235a-4079-a089-43e1b93aef9f" containerID="9f4ec022a951e769e06c755c5e3ee29d2cdb4bf72c3e18edaf341e5d425f9888" exitCode=0 Nov 22 05:10:06 crc kubenswrapper[4818]: I1122 05:10:06.078189 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"69f72526-235a-4079-a089-43e1b93aef9f","Type":"ContainerDied","Data":"9f4ec022a951e769e06c755c5e3ee29d2cdb4bf72c3e18edaf341e5d425f9888"} Nov 22 05:10:06 crc kubenswrapper[4818]: I1122 05:10:06.086132 4818 generic.go:334] "Generic (PLEG): container finished" podID="67e09220-fd61-4785-9de7-90b3ffaf0157" containerID="7b521d1554430e15d45e81217bcbedffcbf9f91f255aee93a639d50189365d64" exitCode=0 Nov 22 05:10:06 crc kubenswrapper[4818]: I1122 05:10:06.086175 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"67e09220-fd61-4785-9de7-90b3ffaf0157","Type":"ContainerDied","Data":"7b521d1554430e15d45e81217bcbedffcbf9f91f255aee93a639d50189365d64"} Nov 22 05:10:06 crc kubenswrapper[4818]: I1122 05:10:06.565281 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-jgq5d" podUID="7dec9ecf-e438-4334-a20e-757ce6fff445" containerName="ovn-controller" probeResult="failure" output=< Nov 22 05:10:06 crc kubenswrapper[4818]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 22 05:10:06 crc kubenswrapper[4818]: > Nov 22 05:10:06 crc kubenswrapper[4818]: I1122 05:10:06.574868 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:10:06 crc kubenswrapper[4818]: I1122 05:10:06.582273 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-kpdh5" Nov 22 05:10:06 crc kubenswrapper[4818]: I1122 05:10:06.801581 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jgq5d-config-nbpkh"] Nov 22 05:10:06 crc kubenswrapper[4818]: I1122 05:10:06.802842 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:06 crc kubenswrapper[4818]: I1122 05:10:06.806864 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jgq5d-config-nbpkh"] Nov 22 05:10:06 crc kubenswrapper[4818]: I1122 05:10:06.808903 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 22 05:10:06 crc kubenswrapper[4818]: I1122 05:10:06.969742 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/82320320-0538-4eef-9667-db96bcc7cb3e-var-run-ovn\") pod \"ovn-controller-jgq5d-config-nbpkh\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:06 crc kubenswrapper[4818]: I1122 05:10:06.969823 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z72jc\" (UniqueName: \"kubernetes.io/projected/82320320-0538-4eef-9667-db96bcc7cb3e-kube-api-access-z72jc\") pod \"ovn-controller-jgq5d-config-nbpkh\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:06 crc kubenswrapper[4818]: I1122 05:10:06.969842 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/82320320-0538-4eef-9667-db96bcc7cb3e-var-log-ovn\") pod \"ovn-controller-jgq5d-config-nbpkh\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:06 crc kubenswrapper[4818]: I1122 05:10:06.969896 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/82320320-0538-4eef-9667-db96bcc7cb3e-additional-scripts\") pod \"ovn-controller-jgq5d-config-nbpkh\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:06 crc kubenswrapper[4818]: I1122 05:10:06.969922 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82320320-0538-4eef-9667-db96bcc7cb3e-scripts\") pod \"ovn-controller-jgq5d-config-nbpkh\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:06 crc kubenswrapper[4818]: I1122 05:10:06.969951 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/82320320-0538-4eef-9667-db96bcc7cb3e-var-run\") pod \"ovn-controller-jgq5d-config-nbpkh\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:07 crc kubenswrapper[4818]: I1122 05:10:07.071158 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/82320320-0538-4eef-9667-db96bcc7cb3e-var-run-ovn\") pod \"ovn-controller-jgq5d-config-nbpkh\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:07 crc kubenswrapper[4818]: I1122 05:10:07.071330 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z72jc\" (UniqueName: \"kubernetes.io/projected/82320320-0538-4eef-9667-db96bcc7cb3e-kube-api-access-z72jc\") pod \"ovn-controller-jgq5d-config-nbpkh\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:07 crc kubenswrapper[4818]: I1122 05:10:07.071369 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/82320320-0538-4eef-9667-db96bcc7cb3e-var-log-ovn\") pod \"ovn-controller-jgq5d-config-nbpkh\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:07 crc kubenswrapper[4818]: I1122 05:10:07.071403 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/82320320-0538-4eef-9667-db96bcc7cb3e-additional-scripts\") pod \"ovn-controller-jgq5d-config-nbpkh\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:07 crc kubenswrapper[4818]: I1122 05:10:07.071457 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82320320-0538-4eef-9667-db96bcc7cb3e-scripts\") pod \"ovn-controller-jgq5d-config-nbpkh\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:07 crc kubenswrapper[4818]: I1122 05:10:07.071499 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/82320320-0538-4eef-9667-db96bcc7cb3e-var-run\") pod \"ovn-controller-jgq5d-config-nbpkh\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:07 crc kubenswrapper[4818]: I1122 05:10:07.071978 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/82320320-0538-4eef-9667-db96bcc7cb3e-var-run\") pod \"ovn-controller-jgq5d-config-nbpkh\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:07 crc kubenswrapper[4818]: I1122 05:10:07.072074 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/82320320-0538-4eef-9667-db96bcc7cb3e-var-run-ovn\") pod \"ovn-controller-jgq5d-config-nbpkh\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:07 crc kubenswrapper[4818]: I1122 05:10:07.072621 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/82320320-0538-4eef-9667-db96bcc7cb3e-var-log-ovn\") pod \"ovn-controller-jgq5d-config-nbpkh\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:07 crc kubenswrapper[4818]: I1122 05:10:07.073659 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/82320320-0538-4eef-9667-db96bcc7cb3e-additional-scripts\") pod \"ovn-controller-jgq5d-config-nbpkh\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:07 crc kubenswrapper[4818]: I1122 05:10:07.077333 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82320320-0538-4eef-9667-db96bcc7cb3e-scripts\") pod \"ovn-controller-jgq5d-config-nbpkh\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:07 crc kubenswrapper[4818]: I1122 05:10:07.094145 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z72jc\" (UniqueName: \"kubernetes.io/projected/82320320-0538-4eef-9667-db96bcc7cb3e-kube-api-access-z72jc\") pod \"ovn-controller-jgq5d-config-nbpkh\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:07 crc kubenswrapper[4818]: I1122 05:10:07.128078 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:11 crc kubenswrapper[4818]: I1122 05:10:11.542673 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-jgq5d" podUID="7dec9ecf-e438-4334-a20e-757ce6fff445" containerName="ovn-controller" probeResult="failure" output=< Nov 22 05:10:11 crc kubenswrapper[4818]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 22 05:10:11 crc kubenswrapper[4818]: > Nov 22 05:10:16 crc kubenswrapper[4818]: E1122 05:10:16.535823 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Nov 22 05:10:16 crc kubenswrapper[4818]: E1122 05:10:16.536632 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-br247,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-jpkvr_openstack(486177c7-b0c3-45ba-a18c-79509154c9df): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 05:10:16 crc kubenswrapper[4818]: E1122 05:10:16.537815 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-jpkvr" podUID="486177c7-b0c3-45ba-a18c-79509154c9df" Nov 22 05:10:16 crc kubenswrapper[4818]: I1122 05:10:16.554014 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-jgq5d" podUID="7dec9ecf-e438-4334-a20e-757ce6fff445" containerName="ovn-controller" probeResult="failure" output=< Nov 22 05:10:16 crc kubenswrapper[4818]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 22 05:10:16 crc kubenswrapper[4818]: > Nov 22 05:10:16 crc kubenswrapper[4818]: I1122 05:10:16.963304 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jgq5d-config-nbpkh"] Nov 22 05:10:17 crc kubenswrapper[4818]: I1122 05:10:17.183705 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jgq5d-config-nbpkh" event={"ID":"82320320-0538-4eef-9667-db96bcc7cb3e","Type":"ContainerStarted","Data":"154dfbecc227a0d198dae7f660f0fd5a7f5e5e6215167e09af810f581f6b25c3"} Nov 22 05:10:17 crc kubenswrapper[4818]: I1122 05:10:17.186903 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"69f72526-235a-4079-a089-43e1b93aef9f","Type":"ContainerStarted","Data":"c51c9537d0a14d91d8b78ef778dfb2df074295aaa1b7aea71a8070be82aeb023"} Nov 22 05:10:17 crc kubenswrapper[4818]: I1122 05:10:17.187142 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 22 05:10:17 crc kubenswrapper[4818]: I1122 05:10:17.189931 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"67e09220-fd61-4785-9de7-90b3ffaf0157","Type":"ContainerStarted","Data":"2011cd97d070b8a754b3acec1fbdf7b22170264b4f513d61217568bcbbd91c82"} Nov 22 05:10:17 crc kubenswrapper[4818]: I1122 05:10:17.190538 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:10:17 crc kubenswrapper[4818]: E1122 05:10:17.191834 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-jpkvr" podUID="486177c7-b0c3-45ba-a18c-79509154c9df" Nov 22 05:10:17 crc kubenswrapper[4818]: I1122 05:10:17.209753 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=72.635995488 podStartE2EDuration="1m22.209733121s" podCreationTimestamp="2025-11-22 05:08:55 +0000 UTC" firstStartedPulling="2025-11-22 05:09:22.430972645 +0000 UTC m=+1315.005389172" lastFinishedPulling="2025-11-22 05:09:32.004710278 +0000 UTC m=+1324.579126805" observedRunningTime="2025-11-22 05:10:17.205794326 +0000 UTC m=+1369.780210863" watchObservedRunningTime="2025-11-22 05:10:17.209733121 +0000 UTC m=+1369.784149648" Nov 22 05:10:17 crc kubenswrapper[4818]: I1122 05:10:17.229846 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=72.654334395 podStartE2EDuration="1m22.229826224s" podCreationTimestamp="2025-11-22 05:08:55 +0000 UTC" firstStartedPulling="2025-11-22 05:09:22.215892269 +0000 UTC m=+1314.790308796" lastFinishedPulling="2025-11-22 05:09:31.791384098 +0000 UTC m=+1324.365800625" observedRunningTime="2025-11-22 05:10:17.22476816 +0000 UTC m=+1369.799184707" watchObservedRunningTime="2025-11-22 05:10:17.229826224 +0000 UTC m=+1369.804242751" Nov 22 05:10:18 crc kubenswrapper[4818]: I1122 05:10:18.198022 4818 generic.go:334] "Generic (PLEG): container finished" podID="82320320-0538-4eef-9667-db96bcc7cb3e" containerID="d487a8c5d84ba9f8f05636aff782142b966c55d4f278ab53dec50fb2b1a1a854" exitCode=0 Nov 22 05:10:18 crc kubenswrapper[4818]: I1122 05:10:18.198074 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jgq5d-config-nbpkh" event={"ID":"82320320-0538-4eef-9667-db96bcc7cb3e","Type":"ContainerDied","Data":"d487a8c5d84ba9f8f05636aff782142b966c55d4f278ab53dec50fb2b1a1a854"} Nov 22 05:10:19 crc kubenswrapper[4818]: I1122 05:10:19.492050 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:19 crc kubenswrapper[4818]: I1122 05:10:19.593535 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82320320-0538-4eef-9667-db96bcc7cb3e-scripts\") pod \"82320320-0538-4eef-9667-db96bcc7cb3e\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " Nov 22 05:10:19 crc kubenswrapper[4818]: I1122 05:10:19.593609 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/82320320-0538-4eef-9667-db96bcc7cb3e-var-run\") pod \"82320320-0538-4eef-9667-db96bcc7cb3e\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " Nov 22 05:10:19 crc kubenswrapper[4818]: I1122 05:10:19.593641 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/82320320-0538-4eef-9667-db96bcc7cb3e-var-log-ovn\") pod \"82320320-0538-4eef-9667-db96bcc7cb3e\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " Nov 22 05:10:19 crc kubenswrapper[4818]: I1122 05:10:19.593672 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z72jc\" (UniqueName: \"kubernetes.io/projected/82320320-0538-4eef-9667-db96bcc7cb3e-kube-api-access-z72jc\") pod \"82320320-0538-4eef-9667-db96bcc7cb3e\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " Nov 22 05:10:19 crc kubenswrapper[4818]: I1122 05:10:19.593728 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/82320320-0538-4eef-9667-db96bcc7cb3e-var-run" (OuterVolumeSpecName: "var-run") pod "82320320-0538-4eef-9667-db96bcc7cb3e" (UID: "82320320-0538-4eef-9667-db96bcc7cb3e"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:10:19 crc kubenswrapper[4818]: I1122 05:10:19.593781 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/82320320-0538-4eef-9667-db96bcc7cb3e-var-run-ovn\") pod \"82320320-0538-4eef-9667-db96bcc7cb3e\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " Nov 22 05:10:19 crc kubenswrapper[4818]: I1122 05:10:19.593813 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/82320320-0538-4eef-9667-db96bcc7cb3e-additional-scripts\") pod \"82320320-0538-4eef-9667-db96bcc7cb3e\" (UID: \"82320320-0538-4eef-9667-db96bcc7cb3e\") " Nov 22 05:10:19 crc kubenswrapper[4818]: I1122 05:10:19.593738 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/82320320-0538-4eef-9667-db96bcc7cb3e-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "82320320-0538-4eef-9667-db96bcc7cb3e" (UID: "82320320-0538-4eef-9667-db96bcc7cb3e"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:10:19 crc kubenswrapper[4818]: I1122 05:10:19.593842 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/82320320-0538-4eef-9667-db96bcc7cb3e-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "82320320-0538-4eef-9667-db96bcc7cb3e" (UID: "82320320-0538-4eef-9667-db96bcc7cb3e"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:10:19 crc kubenswrapper[4818]: I1122 05:10:19.594189 4818 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/82320320-0538-4eef-9667-db96bcc7cb3e-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:19 crc kubenswrapper[4818]: I1122 05:10:19.594211 4818 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/82320320-0538-4eef-9667-db96bcc7cb3e-var-run\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:19 crc kubenswrapper[4818]: I1122 05:10:19.594224 4818 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/82320320-0538-4eef-9667-db96bcc7cb3e-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:19 crc kubenswrapper[4818]: I1122 05:10:19.594472 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82320320-0538-4eef-9667-db96bcc7cb3e-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "82320320-0538-4eef-9667-db96bcc7cb3e" (UID: "82320320-0538-4eef-9667-db96bcc7cb3e"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:10:19 crc kubenswrapper[4818]: I1122 05:10:19.595021 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82320320-0538-4eef-9667-db96bcc7cb3e-scripts" (OuterVolumeSpecName: "scripts") pod "82320320-0538-4eef-9667-db96bcc7cb3e" (UID: "82320320-0538-4eef-9667-db96bcc7cb3e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:10:19 crc kubenswrapper[4818]: I1122 05:10:19.600320 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82320320-0538-4eef-9667-db96bcc7cb3e-kube-api-access-z72jc" (OuterVolumeSpecName: "kube-api-access-z72jc") pod "82320320-0538-4eef-9667-db96bcc7cb3e" (UID: "82320320-0538-4eef-9667-db96bcc7cb3e"). InnerVolumeSpecName "kube-api-access-z72jc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:10:19 crc kubenswrapper[4818]: I1122 05:10:19.695461 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82320320-0538-4eef-9667-db96bcc7cb3e-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:19 crc kubenswrapper[4818]: I1122 05:10:19.695502 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z72jc\" (UniqueName: \"kubernetes.io/projected/82320320-0538-4eef-9667-db96bcc7cb3e-kube-api-access-z72jc\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:19 crc kubenswrapper[4818]: I1122 05:10:19.695521 4818 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/82320320-0538-4eef-9667-db96bcc7cb3e-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.216916 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jgq5d-config-nbpkh" event={"ID":"82320320-0538-4eef-9667-db96bcc7cb3e","Type":"ContainerDied","Data":"154dfbecc227a0d198dae7f660f0fd5a7f5e5e6215167e09af810f581f6b25c3"} Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.216962 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="154dfbecc227a0d198dae7f660f0fd5a7f5e5e6215167e09af810f581f6b25c3" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.217028 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jgq5d-config-nbpkh" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.597400 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-jgq5d-config-nbpkh"] Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.613078 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-jgq5d-config-nbpkh"] Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.721127 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jgq5d-config-d7dsd"] Nov 22 05:10:20 crc kubenswrapper[4818]: E1122 05:10:20.721482 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82320320-0538-4eef-9667-db96bcc7cb3e" containerName="ovn-config" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.721500 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="82320320-0538-4eef-9667-db96bcc7cb3e" containerName="ovn-config" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.721675 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="82320320-0538-4eef-9667-db96bcc7cb3e" containerName="ovn-config" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.722409 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.724928 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.737976 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jgq5d-config-d7dsd"] Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.813142 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97c69\" (UniqueName: \"kubernetes.io/projected/373ddb82-a021-404b-83c7-de441f06a5cf-kube-api-access-97c69\") pod \"ovn-controller-jgq5d-config-d7dsd\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.813209 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/373ddb82-a021-404b-83c7-de441f06a5cf-var-log-ovn\") pod \"ovn-controller-jgq5d-config-d7dsd\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.813283 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/373ddb82-a021-404b-83c7-de441f06a5cf-scripts\") pod \"ovn-controller-jgq5d-config-d7dsd\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.813398 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/373ddb82-a021-404b-83c7-de441f06a5cf-var-run-ovn\") pod \"ovn-controller-jgq5d-config-d7dsd\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.813436 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/373ddb82-a021-404b-83c7-de441f06a5cf-additional-scripts\") pod \"ovn-controller-jgq5d-config-d7dsd\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.813487 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/373ddb82-a021-404b-83c7-de441f06a5cf-var-run\") pod \"ovn-controller-jgq5d-config-d7dsd\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.915407 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/373ddb82-a021-404b-83c7-de441f06a5cf-var-run\") pod \"ovn-controller-jgq5d-config-d7dsd\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.915768 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/373ddb82-a021-404b-83c7-de441f06a5cf-var-run\") pod \"ovn-controller-jgq5d-config-d7dsd\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.915906 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97c69\" (UniqueName: \"kubernetes.io/projected/373ddb82-a021-404b-83c7-de441f06a5cf-kube-api-access-97c69\") pod \"ovn-controller-jgq5d-config-d7dsd\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.916322 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/373ddb82-a021-404b-83c7-de441f06a5cf-var-log-ovn\") pod \"ovn-controller-jgq5d-config-d7dsd\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.916415 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/373ddb82-a021-404b-83c7-de441f06a5cf-var-log-ovn\") pod \"ovn-controller-jgq5d-config-d7dsd\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.916526 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/373ddb82-a021-404b-83c7-de441f06a5cf-scripts\") pod \"ovn-controller-jgq5d-config-d7dsd\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.916559 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/373ddb82-a021-404b-83c7-de441f06a5cf-var-run-ovn\") pod \"ovn-controller-jgq5d-config-d7dsd\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.918600 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/373ddb82-a021-404b-83c7-de441f06a5cf-additional-scripts\") pod \"ovn-controller-jgq5d-config-d7dsd\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.918547 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/373ddb82-a021-404b-83c7-de441f06a5cf-scripts\") pod \"ovn-controller-jgq5d-config-d7dsd\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.916653 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/373ddb82-a021-404b-83c7-de441f06a5cf-var-run-ovn\") pod \"ovn-controller-jgq5d-config-d7dsd\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.919159 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/373ddb82-a021-404b-83c7-de441f06a5cf-additional-scripts\") pod \"ovn-controller-jgq5d-config-d7dsd\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:20 crc kubenswrapper[4818]: I1122 05:10:20.939957 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97c69\" (UniqueName: \"kubernetes.io/projected/373ddb82-a021-404b-83c7-de441f06a5cf-kube-api-access-97c69\") pod \"ovn-controller-jgq5d-config-d7dsd\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:21 crc kubenswrapper[4818]: I1122 05:10:21.040390 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:21 crc kubenswrapper[4818]: I1122 05:10:21.265162 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:10:21 crc kubenswrapper[4818]: I1122 05:10:21.265463 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:10:21 crc kubenswrapper[4818]: I1122 05:10:21.527769 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jgq5d-config-d7dsd"] Nov 22 05:10:21 crc kubenswrapper[4818]: I1122 05:10:21.557700 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-jgq5d" Nov 22 05:10:22 crc kubenswrapper[4818]: I1122 05:10:22.247908 4818 generic.go:334] "Generic (PLEG): container finished" podID="373ddb82-a021-404b-83c7-de441f06a5cf" containerID="3cba6a520a9001c55a8f1e00a44043e22e6e8f1ab4eb4d7d10157a2d09ab9561" exitCode=0 Nov 22 05:10:22 crc kubenswrapper[4818]: I1122 05:10:22.248071 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jgq5d-config-d7dsd" event={"ID":"373ddb82-a021-404b-83c7-de441f06a5cf","Type":"ContainerDied","Data":"3cba6a520a9001c55a8f1e00a44043e22e6e8f1ab4eb4d7d10157a2d09ab9561"} Nov 22 05:10:22 crc kubenswrapper[4818]: I1122 05:10:22.248184 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jgq5d-config-d7dsd" event={"ID":"373ddb82-a021-404b-83c7-de441f06a5cf","Type":"ContainerStarted","Data":"f10860efdc2b96ffa61e2cb00ffc64a2f08ccff561de36659285e260d8e9cb9c"} Nov 22 05:10:22 crc kubenswrapper[4818]: I1122 05:10:22.302408 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82320320-0538-4eef-9667-db96bcc7cb3e" path="/var/lib/kubelet/pods/82320320-0538-4eef-9667-db96bcc7cb3e/volumes" Nov 22 05:10:23 crc kubenswrapper[4818]: I1122 05:10:23.551380 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:23 crc kubenswrapper[4818]: I1122 05:10:23.667876 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/373ddb82-a021-404b-83c7-de441f06a5cf-var-log-ovn\") pod \"373ddb82-a021-404b-83c7-de441f06a5cf\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " Nov 22 05:10:23 crc kubenswrapper[4818]: I1122 05:10:23.667942 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/373ddb82-a021-404b-83c7-de441f06a5cf-additional-scripts\") pod \"373ddb82-a021-404b-83c7-de441f06a5cf\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " Nov 22 05:10:23 crc kubenswrapper[4818]: I1122 05:10:23.668006 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/373ddb82-a021-404b-83c7-de441f06a5cf-var-run\") pod \"373ddb82-a021-404b-83c7-de441f06a5cf\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " Nov 22 05:10:23 crc kubenswrapper[4818]: I1122 05:10:23.668054 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97c69\" (UniqueName: \"kubernetes.io/projected/373ddb82-a021-404b-83c7-de441f06a5cf-kube-api-access-97c69\") pod \"373ddb82-a021-404b-83c7-de441f06a5cf\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " Nov 22 05:10:23 crc kubenswrapper[4818]: I1122 05:10:23.668095 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/373ddb82-a021-404b-83c7-de441f06a5cf-var-run-ovn\") pod \"373ddb82-a021-404b-83c7-de441f06a5cf\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " Nov 22 05:10:23 crc kubenswrapper[4818]: I1122 05:10:23.668165 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/373ddb82-a021-404b-83c7-de441f06a5cf-scripts\") pod \"373ddb82-a021-404b-83c7-de441f06a5cf\" (UID: \"373ddb82-a021-404b-83c7-de441f06a5cf\") " Nov 22 05:10:23 crc kubenswrapper[4818]: I1122 05:10:23.668266 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/373ddb82-a021-404b-83c7-de441f06a5cf-var-run" (OuterVolumeSpecName: "var-run") pod "373ddb82-a021-404b-83c7-de441f06a5cf" (UID: "373ddb82-a021-404b-83c7-de441f06a5cf"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:10:23 crc kubenswrapper[4818]: I1122 05:10:23.668330 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/373ddb82-a021-404b-83c7-de441f06a5cf-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "373ddb82-a021-404b-83c7-de441f06a5cf" (UID: "373ddb82-a021-404b-83c7-de441f06a5cf"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:10:23 crc kubenswrapper[4818]: I1122 05:10:23.668523 4818 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/373ddb82-a021-404b-83c7-de441f06a5cf-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:23 crc kubenswrapper[4818]: I1122 05:10:23.668540 4818 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/373ddb82-a021-404b-83c7-de441f06a5cf-var-run\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:23 crc kubenswrapper[4818]: I1122 05:10:23.669183 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/373ddb82-a021-404b-83c7-de441f06a5cf-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "373ddb82-a021-404b-83c7-de441f06a5cf" (UID: "373ddb82-a021-404b-83c7-de441f06a5cf"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:10:23 crc kubenswrapper[4818]: I1122 05:10:23.669481 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/373ddb82-a021-404b-83c7-de441f06a5cf-scripts" (OuterVolumeSpecName: "scripts") pod "373ddb82-a021-404b-83c7-de441f06a5cf" (UID: "373ddb82-a021-404b-83c7-de441f06a5cf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:10:23 crc kubenswrapper[4818]: I1122 05:10:23.669524 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/373ddb82-a021-404b-83c7-de441f06a5cf-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "373ddb82-a021-404b-83c7-de441f06a5cf" (UID: "373ddb82-a021-404b-83c7-de441f06a5cf"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:10:23 crc kubenswrapper[4818]: I1122 05:10:23.676434 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/373ddb82-a021-404b-83c7-de441f06a5cf-kube-api-access-97c69" (OuterVolumeSpecName: "kube-api-access-97c69") pod "373ddb82-a021-404b-83c7-de441f06a5cf" (UID: "373ddb82-a021-404b-83c7-de441f06a5cf"). InnerVolumeSpecName "kube-api-access-97c69". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:10:23 crc kubenswrapper[4818]: I1122 05:10:23.770807 4818 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/373ddb82-a021-404b-83c7-de441f06a5cf-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:23 crc kubenswrapper[4818]: I1122 05:10:23.770862 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/373ddb82-a021-404b-83c7-de441f06a5cf-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:23 crc kubenswrapper[4818]: I1122 05:10:23.770884 4818 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/373ddb82-a021-404b-83c7-de441f06a5cf-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:23 crc kubenswrapper[4818]: I1122 05:10:23.770904 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97c69\" (UniqueName: \"kubernetes.io/projected/373ddb82-a021-404b-83c7-de441f06a5cf-kube-api-access-97c69\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:24 crc kubenswrapper[4818]: I1122 05:10:24.268232 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jgq5d-config-d7dsd" event={"ID":"373ddb82-a021-404b-83c7-de441f06a5cf","Type":"ContainerDied","Data":"f10860efdc2b96ffa61e2cb00ffc64a2f08ccff561de36659285e260d8e9cb9c"} Nov 22 05:10:24 crc kubenswrapper[4818]: I1122 05:10:24.268381 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f10860efdc2b96ffa61e2cb00ffc64a2f08ccff561de36659285e260d8e9cb9c" Nov 22 05:10:24 crc kubenswrapper[4818]: I1122 05:10:24.268470 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jgq5d-config-d7dsd" Nov 22 05:10:24 crc kubenswrapper[4818]: I1122 05:10:24.633387 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-jgq5d-config-d7dsd"] Nov 22 05:10:24 crc kubenswrapper[4818]: I1122 05:10:24.638695 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-jgq5d-config-d7dsd"] Nov 22 05:10:26 crc kubenswrapper[4818]: I1122 05:10:26.305888 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="373ddb82-a021-404b-83c7-de441f06a5cf" path="/var/lib/kubelet/pods/373ddb82-a021-404b-83c7-de441f06a5cf/volumes" Nov 22 05:10:26 crc kubenswrapper[4818]: I1122 05:10:26.664525 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 22 05:10:26 crc kubenswrapper[4818]: I1122 05:10:26.922314 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-zktzs"] Nov 22 05:10:26 crc kubenswrapper[4818]: E1122 05:10:26.922741 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="373ddb82-a021-404b-83c7-de441f06a5cf" containerName="ovn-config" Nov 22 05:10:26 crc kubenswrapper[4818]: I1122 05:10:26.922764 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="373ddb82-a021-404b-83c7-de441f06a5cf" containerName="ovn-config" Nov 22 05:10:26 crc kubenswrapper[4818]: I1122 05:10:26.922980 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="373ddb82-a021-404b-83c7-de441f06a5cf" containerName="ovn-config" Nov 22 05:10:26 crc kubenswrapper[4818]: I1122 05:10:26.923841 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-zktzs" Nov 22 05:10:26 crc kubenswrapper[4818]: I1122 05:10:26.933269 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-zktzs"] Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.016051 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-h7nxw"] Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.017342 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-h7nxw" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.027776 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9ad29a8-c449-4ac3-a30d-64979050a1b0-operator-scripts\") pod \"cinder-db-create-zktzs\" (UID: \"a9ad29a8-c449-4ac3-a30d-64979050a1b0\") " pod="openstack/cinder-db-create-zktzs" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.027843 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j4s4\" (UniqueName: \"kubernetes.io/projected/a9ad29a8-c449-4ac3-a30d-64979050a1b0-kube-api-access-4j4s4\") pod \"cinder-db-create-zktzs\" (UID: \"a9ad29a8-c449-4ac3-a30d-64979050a1b0\") " pod="openstack/cinder-db-create-zktzs" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.030416 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.045840 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-d2da-account-create-zw6nl"] Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.047434 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d2da-account-create-zw6nl" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.049849 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.066445 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-d2da-account-create-zw6nl"] Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.077560 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-h7nxw"] Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.135932 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60fc3ace-6591-4551-96ba-a746eacf5609-operator-scripts\") pod \"barbican-db-create-h7nxw\" (UID: \"60fc3ace-6591-4551-96ba-a746eacf5609\") " pod="openstack/barbican-db-create-h7nxw" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.136016 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9ad29a8-c449-4ac3-a30d-64979050a1b0-operator-scripts\") pod \"cinder-db-create-zktzs\" (UID: \"a9ad29a8-c449-4ac3-a30d-64979050a1b0\") " pod="openstack/cinder-db-create-zktzs" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.136198 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j4s4\" (UniqueName: \"kubernetes.io/projected/a9ad29a8-c449-4ac3-a30d-64979050a1b0-kube-api-access-4j4s4\") pod \"cinder-db-create-zktzs\" (UID: \"a9ad29a8-c449-4ac3-a30d-64979050a1b0\") " pod="openstack/cinder-db-create-zktzs" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.136317 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2v6l9\" (UniqueName: \"kubernetes.io/projected/d879cd8e-9523-47ac-acfc-75aa7fde0bc0-kube-api-access-2v6l9\") pod \"barbican-d2da-account-create-zw6nl\" (UID: \"d879cd8e-9523-47ac-acfc-75aa7fde0bc0\") " pod="openstack/barbican-d2da-account-create-zw6nl" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.136457 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krnvd\" (UniqueName: \"kubernetes.io/projected/60fc3ace-6591-4551-96ba-a746eacf5609-kube-api-access-krnvd\") pod \"barbican-db-create-h7nxw\" (UID: \"60fc3ace-6591-4551-96ba-a746eacf5609\") " pod="openstack/barbican-db-create-h7nxw" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.136510 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d879cd8e-9523-47ac-acfc-75aa7fde0bc0-operator-scripts\") pod \"barbican-d2da-account-create-zw6nl\" (UID: \"d879cd8e-9523-47ac-acfc-75aa7fde0bc0\") " pod="openstack/barbican-d2da-account-create-zw6nl" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.136860 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9ad29a8-c449-4ac3-a30d-64979050a1b0-operator-scripts\") pod \"cinder-db-create-zktzs\" (UID: \"a9ad29a8-c449-4ac3-a30d-64979050a1b0\") " pod="openstack/cinder-db-create-zktzs" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.154856 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-ea58-account-create-rxpn5"] Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.156153 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ea58-account-create-rxpn5" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.159414 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.163614 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ea58-account-create-rxpn5"] Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.164412 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j4s4\" (UniqueName: \"kubernetes.io/projected/a9ad29a8-c449-4ac3-a30d-64979050a1b0-kube-api-access-4j4s4\") pod \"cinder-db-create-zktzs\" (UID: \"a9ad29a8-c449-4ac3-a30d-64979050a1b0\") " pod="openstack/cinder-db-create-zktzs" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.238132 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60fc3ace-6591-4551-96ba-a746eacf5609-operator-scripts\") pod \"barbican-db-create-h7nxw\" (UID: \"60fc3ace-6591-4551-96ba-a746eacf5609\") " pod="openstack/barbican-db-create-h7nxw" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.238186 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc03cbc0-add4-4e90-a74a-edc369ee88b9-operator-scripts\") pod \"cinder-ea58-account-create-rxpn5\" (UID: \"cc03cbc0-add4-4e90-a74a-edc369ee88b9\") " pod="openstack/cinder-ea58-account-create-rxpn5" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.238242 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2v6l9\" (UniqueName: \"kubernetes.io/projected/d879cd8e-9523-47ac-acfc-75aa7fde0bc0-kube-api-access-2v6l9\") pod \"barbican-d2da-account-create-zw6nl\" (UID: \"d879cd8e-9523-47ac-acfc-75aa7fde0bc0\") " pod="openstack/barbican-d2da-account-create-zw6nl" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.238303 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krnvd\" (UniqueName: \"kubernetes.io/projected/60fc3ace-6591-4551-96ba-a746eacf5609-kube-api-access-krnvd\") pod \"barbican-db-create-h7nxw\" (UID: \"60fc3ace-6591-4551-96ba-a746eacf5609\") " pod="openstack/barbican-db-create-h7nxw" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.238324 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc8pr\" (UniqueName: \"kubernetes.io/projected/cc03cbc0-add4-4e90-a74a-edc369ee88b9-kube-api-access-lc8pr\") pod \"cinder-ea58-account-create-rxpn5\" (UID: \"cc03cbc0-add4-4e90-a74a-edc369ee88b9\") " pod="openstack/cinder-ea58-account-create-rxpn5" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.238342 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d879cd8e-9523-47ac-acfc-75aa7fde0bc0-operator-scripts\") pod \"barbican-d2da-account-create-zw6nl\" (UID: \"d879cd8e-9523-47ac-acfc-75aa7fde0bc0\") " pod="openstack/barbican-d2da-account-create-zw6nl" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.238857 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60fc3ace-6591-4551-96ba-a746eacf5609-operator-scripts\") pod \"barbican-db-create-h7nxw\" (UID: \"60fc3ace-6591-4551-96ba-a746eacf5609\") " pod="openstack/barbican-db-create-h7nxw" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.239007 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d879cd8e-9523-47ac-acfc-75aa7fde0bc0-operator-scripts\") pod \"barbican-d2da-account-create-zw6nl\" (UID: \"d879cd8e-9523-47ac-acfc-75aa7fde0bc0\") " pod="openstack/barbican-d2da-account-create-zw6nl" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.241855 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-zktzs" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.258782 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krnvd\" (UniqueName: \"kubernetes.io/projected/60fc3ace-6591-4551-96ba-a746eacf5609-kube-api-access-krnvd\") pod \"barbican-db-create-h7nxw\" (UID: \"60fc3ace-6591-4551-96ba-a746eacf5609\") " pod="openstack/barbican-db-create-h7nxw" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.266296 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2v6l9\" (UniqueName: \"kubernetes.io/projected/d879cd8e-9523-47ac-acfc-75aa7fde0bc0-kube-api-access-2v6l9\") pod \"barbican-d2da-account-create-zw6nl\" (UID: \"d879cd8e-9523-47ac-acfc-75aa7fde0bc0\") " pod="openstack/barbican-d2da-account-create-zw6nl" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.327927 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-4ggzd"] Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.329202 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4ggzd" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.331345 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-h7nxw" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.341507 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc03cbc0-add4-4e90-a74a-edc369ee88b9-operator-scripts\") pod \"cinder-ea58-account-create-rxpn5\" (UID: \"cc03cbc0-add4-4e90-a74a-edc369ee88b9\") " pod="openstack/cinder-ea58-account-create-rxpn5" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.341651 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc8pr\" (UniqueName: \"kubernetes.io/projected/cc03cbc0-add4-4e90-a74a-edc369ee88b9-kube-api-access-lc8pr\") pod \"cinder-ea58-account-create-rxpn5\" (UID: \"cc03cbc0-add4-4e90-a74a-edc369ee88b9\") " pod="openstack/cinder-ea58-account-create-rxpn5" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.343024 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc03cbc0-add4-4e90-a74a-edc369ee88b9-operator-scripts\") pod \"cinder-ea58-account-create-rxpn5\" (UID: \"cc03cbc0-add4-4e90-a74a-edc369ee88b9\") " pod="openstack/cinder-ea58-account-create-rxpn5" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.347982 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-4ggzd"] Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.360389 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d2da-account-create-zw6nl" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.418790 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc8pr\" (UniqueName: \"kubernetes.io/projected/cc03cbc0-add4-4e90-a74a-edc369ee88b9-kube-api-access-lc8pr\") pod \"cinder-ea58-account-create-rxpn5\" (UID: \"cc03cbc0-add4-4e90-a74a-edc369ee88b9\") " pod="openstack/cinder-ea58-account-create-rxpn5" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.442695 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqsqj\" (UniqueName: \"kubernetes.io/projected/147cf615-2bd3-4db4-8d5a-40711c2b8f92-kube-api-access-pqsqj\") pod \"neutron-db-create-4ggzd\" (UID: \"147cf615-2bd3-4db4-8d5a-40711c2b8f92\") " pod="openstack/neutron-db-create-4ggzd" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.442740 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/147cf615-2bd3-4db4-8d5a-40711c2b8f92-operator-scripts\") pod \"neutron-db-create-4ggzd\" (UID: \"147cf615-2bd3-4db4-8d5a-40711c2b8f92\") " pod="openstack/neutron-db-create-4ggzd" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.455148 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-xdpgh"] Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.456735 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xdpgh" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.459368 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.459562 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.459815 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-tds7x" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.459988 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.486811 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xdpgh"] Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.495860 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-bacb-account-create-grbqp"] Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.497310 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bacb-account-create-grbqp" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.499350 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.509870 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ea58-account-create-rxpn5" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.511648 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-bacb-account-create-grbqp"] Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.544021 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86b3ecec-fc53-4d2d-843b-ab24ea43d82a-operator-scripts\") pod \"neutron-bacb-account-create-grbqp\" (UID: \"86b3ecec-fc53-4d2d-843b-ab24ea43d82a\") " pod="openstack/neutron-bacb-account-create-grbqp" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.544070 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpzg6\" (UniqueName: \"kubernetes.io/projected/06f95278-ba08-4ccd-af26-48c0c06157e7-kube-api-access-bpzg6\") pod \"keystone-db-sync-xdpgh\" (UID: \"06f95278-ba08-4ccd-af26-48c0c06157e7\") " pod="openstack/keystone-db-sync-xdpgh" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.544092 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06f95278-ba08-4ccd-af26-48c0c06157e7-config-data\") pod \"keystone-db-sync-xdpgh\" (UID: \"06f95278-ba08-4ccd-af26-48c0c06157e7\") " pod="openstack/keystone-db-sync-xdpgh" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.544116 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqsqj\" (UniqueName: \"kubernetes.io/projected/147cf615-2bd3-4db4-8d5a-40711c2b8f92-kube-api-access-pqsqj\") pod \"neutron-db-create-4ggzd\" (UID: \"147cf615-2bd3-4db4-8d5a-40711c2b8f92\") " pod="openstack/neutron-db-create-4ggzd" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.544139 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw6cm\" (UniqueName: \"kubernetes.io/projected/86b3ecec-fc53-4d2d-843b-ab24ea43d82a-kube-api-access-fw6cm\") pod \"neutron-bacb-account-create-grbqp\" (UID: \"86b3ecec-fc53-4d2d-843b-ab24ea43d82a\") " pod="openstack/neutron-bacb-account-create-grbqp" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.544156 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/147cf615-2bd3-4db4-8d5a-40711c2b8f92-operator-scripts\") pod \"neutron-db-create-4ggzd\" (UID: \"147cf615-2bd3-4db4-8d5a-40711c2b8f92\") " pod="openstack/neutron-db-create-4ggzd" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.544181 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f95278-ba08-4ccd-af26-48c0c06157e7-combined-ca-bundle\") pod \"keystone-db-sync-xdpgh\" (UID: \"06f95278-ba08-4ccd-af26-48c0c06157e7\") " pod="openstack/keystone-db-sync-xdpgh" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.545845 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/147cf615-2bd3-4db4-8d5a-40711c2b8f92-operator-scripts\") pod \"neutron-db-create-4ggzd\" (UID: \"147cf615-2bd3-4db4-8d5a-40711c2b8f92\") " pod="openstack/neutron-db-create-4ggzd" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.564581 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqsqj\" (UniqueName: \"kubernetes.io/projected/147cf615-2bd3-4db4-8d5a-40711c2b8f92-kube-api-access-pqsqj\") pod \"neutron-db-create-4ggzd\" (UID: \"147cf615-2bd3-4db4-8d5a-40711c2b8f92\") " pod="openstack/neutron-db-create-4ggzd" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.646754 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86b3ecec-fc53-4d2d-843b-ab24ea43d82a-operator-scripts\") pod \"neutron-bacb-account-create-grbqp\" (UID: \"86b3ecec-fc53-4d2d-843b-ab24ea43d82a\") " pod="openstack/neutron-bacb-account-create-grbqp" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.646831 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpzg6\" (UniqueName: \"kubernetes.io/projected/06f95278-ba08-4ccd-af26-48c0c06157e7-kube-api-access-bpzg6\") pod \"keystone-db-sync-xdpgh\" (UID: \"06f95278-ba08-4ccd-af26-48c0c06157e7\") " pod="openstack/keystone-db-sync-xdpgh" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.646865 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06f95278-ba08-4ccd-af26-48c0c06157e7-config-data\") pod \"keystone-db-sync-xdpgh\" (UID: \"06f95278-ba08-4ccd-af26-48c0c06157e7\") " pod="openstack/keystone-db-sync-xdpgh" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.646906 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw6cm\" (UniqueName: \"kubernetes.io/projected/86b3ecec-fc53-4d2d-843b-ab24ea43d82a-kube-api-access-fw6cm\") pod \"neutron-bacb-account-create-grbqp\" (UID: \"86b3ecec-fc53-4d2d-843b-ab24ea43d82a\") " pod="openstack/neutron-bacb-account-create-grbqp" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.646943 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f95278-ba08-4ccd-af26-48c0c06157e7-combined-ca-bundle\") pod \"keystone-db-sync-xdpgh\" (UID: \"06f95278-ba08-4ccd-af26-48c0c06157e7\") " pod="openstack/keystone-db-sync-xdpgh" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.648317 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86b3ecec-fc53-4d2d-843b-ab24ea43d82a-operator-scripts\") pod \"neutron-bacb-account-create-grbqp\" (UID: \"86b3ecec-fc53-4d2d-843b-ab24ea43d82a\") " pod="openstack/neutron-bacb-account-create-grbqp" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.651382 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06f95278-ba08-4ccd-af26-48c0c06157e7-config-data\") pod \"keystone-db-sync-xdpgh\" (UID: \"06f95278-ba08-4ccd-af26-48c0c06157e7\") " pod="openstack/keystone-db-sync-xdpgh" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.652962 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f95278-ba08-4ccd-af26-48c0c06157e7-combined-ca-bundle\") pod \"keystone-db-sync-xdpgh\" (UID: \"06f95278-ba08-4ccd-af26-48c0c06157e7\") " pod="openstack/keystone-db-sync-xdpgh" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.665983 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpzg6\" (UniqueName: \"kubernetes.io/projected/06f95278-ba08-4ccd-af26-48c0c06157e7-kube-api-access-bpzg6\") pod \"keystone-db-sync-xdpgh\" (UID: \"06f95278-ba08-4ccd-af26-48c0c06157e7\") " pod="openstack/keystone-db-sync-xdpgh" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.669732 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw6cm\" (UniqueName: \"kubernetes.io/projected/86b3ecec-fc53-4d2d-843b-ab24ea43d82a-kube-api-access-fw6cm\") pod \"neutron-bacb-account-create-grbqp\" (UID: \"86b3ecec-fc53-4d2d-843b-ab24ea43d82a\") " pod="openstack/neutron-bacb-account-create-grbqp" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.798931 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4ggzd" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.819606 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xdpgh" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.829758 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bacb-account-create-grbqp" Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.847789 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-zktzs"] Nov 22 05:10:27 crc kubenswrapper[4818]: W1122 05:10:27.856308 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9ad29a8_c449_4ac3_a30d_64979050a1b0.slice/crio-795922aae444354441584f6b0fd8d1d67991246eb508bcb3475f14902827d5fe WatchSource:0}: Error finding container 795922aae444354441584f6b0fd8d1d67991246eb508bcb3475f14902827d5fe: Status 404 returned error can't find the container with id 795922aae444354441584f6b0fd8d1d67991246eb508bcb3475f14902827d5fe Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.935063 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-h7nxw"] Nov 22 05:10:27 crc kubenswrapper[4818]: I1122 05:10:27.967603 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-d2da-account-create-zw6nl"] Nov 22 05:10:27 crc kubenswrapper[4818]: W1122 05:10:27.991226 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd879cd8e_9523_47ac_acfc_75aa7fde0bc0.slice/crio-cdeb1d3ba60a18c5cc0b7c368e42e97b186921a5598771086c4f03db98950f96 WatchSource:0}: Error finding container cdeb1d3ba60a18c5cc0b7c368e42e97b186921a5598771086c4f03db98950f96: Status 404 returned error can't find the container with id cdeb1d3ba60a18c5cc0b7c368e42e97b186921a5598771086c4f03db98950f96 Nov 22 05:10:28 crc kubenswrapper[4818]: I1122 05:10:28.042171 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ea58-account-create-rxpn5"] Nov 22 05:10:28 crc kubenswrapper[4818]: I1122 05:10:28.094558 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-4ggzd"] Nov 22 05:10:28 crc kubenswrapper[4818]: I1122 05:10:28.159995 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xdpgh"] Nov 22 05:10:28 crc kubenswrapper[4818]: W1122 05:10:28.171992 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06f95278_ba08_4ccd_af26_48c0c06157e7.slice/crio-3d8a02ed5bd25dfbcb523c84dd32956c77684be759aab7ca584d53c55e3a8c12 WatchSource:0}: Error finding container 3d8a02ed5bd25dfbcb523c84dd32956c77684be759aab7ca584d53c55e3a8c12: Status 404 returned error can't find the container with id 3d8a02ed5bd25dfbcb523c84dd32956c77684be759aab7ca584d53c55e3a8c12 Nov 22 05:10:28 crc kubenswrapper[4818]: W1122 05:10:28.221372 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86b3ecec_fc53_4d2d_843b_ab24ea43d82a.slice/crio-4fa7c10b78e723112464de0172da9b1275c28eba762488f99de2207961bcf3df WatchSource:0}: Error finding container 4fa7c10b78e723112464de0172da9b1275c28eba762488f99de2207961bcf3df: Status 404 returned error can't find the container with id 4fa7c10b78e723112464de0172da9b1275c28eba762488f99de2207961bcf3df Nov 22 05:10:28 crc kubenswrapper[4818]: I1122 05:10:28.222485 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-bacb-account-create-grbqp"] Nov 22 05:10:28 crc kubenswrapper[4818]: I1122 05:10:28.227982 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 22 05:10:28 crc kubenswrapper[4818]: I1122 05:10:28.311795 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4ggzd" event={"ID":"147cf615-2bd3-4db4-8d5a-40711c2b8f92","Type":"ContainerStarted","Data":"f1b5e7531973d206a6ce88608af9b22441740e7b96e6ea2d7709b55bd14497a5"} Nov 22 05:10:28 crc kubenswrapper[4818]: I1122 05:10:28.313103 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ea58-account-create-rxpn5" event={"ID":"cc03cbc0-add4-4e90-a74a-edc369ee88b9","Type":"ContainerStarted","Data":"df5a1c8d6b1ae68986bb9f844bad82eeab15d6be39ef18278d2a011e7cd6ee5c"} Nov 22 05:10:28 crc kubenswrapper[4818]: I1122 05:10:28.314167 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d2da-account-create-zw6nl" event={"ID":"d879cd8e-9523-47ac-acfc-75aa7fde0bc0","Type":"ContainerStarted","Data":"cdeb1d3ba60a18c5cc0b7c368e42e97b186921a5598771086c4f03db98950f96"} Nov 22 05:10:28 crc kubenswrapper[4818]: I1122 05:10:28.316343 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-zktzs" event={"ID":"a9ad29a8-c449-4ac3-a30d-64979050a1b0","Type":"ContainerStarted","Data":"795922aae444354441584f6b0fd8d1d67991246eb508bcb3475f14902827d5fe"} Nov 22 05:10:28 crc kubenswrapper[4818]: I1122 05:10:28.325265 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xdpgh" event={"ID":"06f95278-ba08-4ccd-af26-48c0c06157e7","Type":"ContainerStarted","Data":"3d8a02ed5bd25dfbcb523c84dd32956c77684be759aab7ca584d53c55e3a8c12"} Nov 22 05:10:28 crc kubenswrapper[4818]: I1122 05:10:28.327341 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bacb-account-create-grbqp" event={"ID":"86b3ecec-fc53-4d2d-843b-ab24ea43d82a","Type":"ContainerStarted","Data":"4fa7c10b78e723112464de0172da9b1275c28eba762488f99de2207961bcf3df"} Nov 22 05:10:28 crc kubenswrapper[4818]: I1122 05:10:28.328453 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-h7nxw" event={"ID":"60fc3ace-6591-4551-96ba-a746eacf5609","Type":"ContainerStarted","Data":"7aa0cc59822d5502830a54c4631611867fd9ea0a42e3dfb912c15432477a1224"} Nov 22 05:10:29 crc kubenswrapper[4818]: I1122 05:10:29.335941 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d2da-account-create-zw6nl" event={"ID":"d879cd8e-9523-47ac-acfc-75aa7fde0bc0","Type":"ContainerStarted","Data":"6a799794c95bebfddd1b6311071d289ac1b4c7c16d72381aa73ebacac14e3b36"} Nov 22 05:10:30 crc kubenswrapper[4818]: I1122 05:10:30.345975 4818 generic.go:334] "Generic (PLEG): container finished" podID="d879cd8e-9523-47ac-acfc-75aa7fde0bc0" containerID="6a799794c95bebfddd1b6311071d289ac1b4c7c16d72381aa73ebacac14e3b36" exitCode=0 Nov 22 05:10:30 crc kubenswrapper[4818]: I1122 05:10:30.346088 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d2da-account-create-zw6nl" event={"ID":"d879cd8e-9523-47ac-acfc-75aa7fde0bc0","Type":"ContainerDied","Data":"6a799794c95bebfddd1b6311071d289ac1b4c7c16d72381aa73ebacac14e3b36"} Nov 22 05:10:30 crc kubenswrapper[4818]: I1122 05:10:30.352081 4818 generic.go:334] "Generic (PLEG): container finished" podID="a9ad29a8-c449-4ac3-a30d-64979050a1b0" containerID="36c4045266d0598dc42d139ad7a2c151990072cc88a96c5cbfcf54fa6fbae720" exitCode=0 Nov 22 05:10:30 crc kubenswrapper[4818]: I1122 05:10:30.352202 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-zktzs" event={"ID":"a9ad29a8-c449-4ac3-a30d-64979050a1b0","Type":"ContainerDied","Data":"36c4045266d0598dc42d139ad7a2c151990072cc88a96c5cbfcf54fa6fbae720"} Nov 22 05:10:30 crc kubenswrapper[4818]: I1122 05:10:30.354009 4818 generic.go:334] "Generic (PLEG): container finished" podID="86b3ecec-fc53-4d2d-843b-ab24ea43d82a" containerID="590bad9f87ba50586a0e48ad3beab578112c10c64bf6bd909d17882368a2103a" exitCode=0 Nov 22 05:10:30 crc kubenswrapper[4818]: I1122 05:10:30.354086 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bacb-account-create-grbqp" event={"ID":"86b3ecec-fc53-4d2d-843b-ab24ea43d82a","Type":"ContainerDied","Data":"590bad9f87ba50586a0e48ad3beab578112c10c64bf6bd909d17882368a2103a"} Nov 22 05:10:30 crc kubenswrapper[4818]: I1122 05:10:30.361558 4818 generic.go:334] "Generic (PLEG): container finished" podID="60fc3ace-6591-4551-96ba-a746eacf5609" containerID="44b20ad92d8fa5547099feca6c4ddb17c309294b0a9936a67b4a6afbf04eddae" exitCode=0 Nov 22 05:10:30 crc kubenswrapper[4818]: I1122 05:10:30.361602 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-h7nxw" event={"ID":"60fc3ace-6591-4551-96ba-a746eacf5609","Type":"ContainerDied","Data":"44b20ad92d8fa5547099feca6c4ddb17c309294b0a9936a67b4a6afbf04eddae"} Nov 22 05:10:30 crc kubenswrapper[4818]: I1122 05:10:30.363385 4818 generic.go:334] "Generic (PLEG): container finished" podID="147cf615-2bd3-4db4-8d5a-40711c2b8f92" containerID="d703a24247a1ed221077b7214cf4452d59d409d2d95b9481af56b9187dd209b4" exitCode=0 Nov 22 05:10:30 crc kubenswrapper[4818]: I1122 05:10:30.363558 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4ggzd" event={"ID":"147cf615-2bd3-4db4-8d5a-40711c2b8f92","Type":"ContainerDied","Data":"d703a24247a1ed221077b7214cf4452d59d409d2d95b9481af56b9187dd209b4"} Nov 22 05:10:30 crc kubenswrapper[4818]: I1122 05:10:30.364990 4818 generic.go:334] "Generic (PLEG): container finished" podID="cc03cbc0-add4-4e90-a74a-edc369ee88b9" containerID="23df6b8419f9081d2d2d35ffa2f9b3a44a1462a282262165b265e12c7befdbc8" exitCode=0 Nov 22 05:10:30 crc kubenswrapper[4818]: I1122 05:10:30.365042 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ea58-account-create-rxpn5" event={"ID":"cc03cbc0-add4-4e90-a74a-edc369ee88b9","Type":"ContainerDied","Data":"23df6b8419f9081d2d2d35ffa2f9b3a44a1462a282262165b265e12c7befdbc8"} Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.430290 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d2da-account-create-zw6nl" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.438705 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4ggzd" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.441126 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-zktzs" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.447232 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ea58-account-create-rxpn5" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.461892 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-h7nxw" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.462715 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4ggzd" event={"ID":"147cf615-2bd3-4db4-8d5a-40711c2b8f92","Type":"ContainerDied","Data":"f1b5e7531973d206a6ce88608af9b22441740e7b96e6ea2d7709b55bd14497a5"} Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.462744 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1b5e7531973d206a6ce88608af9b22441740e7b96e6ea2d7709b55bd14497a5" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.462945 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4ggzd" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.471709 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bacb-account-create-grbqp" event={"ID":"86b3ecec-fc53-4d2d-843b-ab24ea43d82a","Type":"ContainerDied","Data":"4fa7c10b78e723112464de0172da9b1275c28eba762488f99de2207961bcf3df"} Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.471773 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fa7c10b78e723112464de0172da9b1275c28eba762488f99de2207961bcf3df" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.485476 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bacb-account-create-grbqp" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.485559 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-h7nxw" event={"ID":"60fc3ace-6591-4551-96ba-a746eacf5609","Type":"ContainerDied","Data":"7aa0cc59822d5502830a54c4631611867fd9ea0a42e3dfb912c15432477a1224"} Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.485604 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7aa0cc59822d5502830a54c4631611867fd9ea0a42e3dfb912c15432477a1224" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.487128 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-h7nxw" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.503127 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ea58-account-create-rxpn5" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.503172 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ea58-account-create-rxpn5" event={"ID":"cc03cbc0-add4-4e90-a74a-edc369ee88b9","Type":"ContainerDied","Data":"df5a1c8d6b1ae68986bb9f844bad82eeab15d6be39ef18278d2a011e7cd6ee5c"} Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.503227 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df5a1c8d6b1ae68986bb9f844bad82eeab15d6be39ef18278d2a011e7cd6ee5c" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.504857 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d2da-account-create-zw6nl" event={"ID":"d879cd8e-9523-47ac-acfc-75aa7fde0bc0","Type":"ContainerDied","Data":"cdeb1d3ba60a18c5cc0b7c368e42e97b186921a5598771086c4f03db98950f96"} Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.505027 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cdeb1d3ba60a18c5cc0b7c368e42e97b186921a5598771086c4f03db98950f96" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.504914 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d2da-account-create-zw6nl" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.506029 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-zktzs" event={"ID":"a9ad29a8-c449-4ac3-a30d-64979050a1b0","Type":"ContainerDied","Data":"795922aae444354441584f6b0fd8d1d67991246eb508bcb3475f14902827d5fe"} Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.506056 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="795922aae444354441584f6b0fd8d1d67991246eb508bcb3475f14902827d5fe" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.506394 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-zktzs" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.568796 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60fc3ace-6591-4551-96ba-a746eacf5609-operator-scripts\") pod \"60fc3ace-6591-4551-96ba-a746eacf5609\" (UID: \"60fc3ace-6591-4551-96ba-a746eacf5609\") " Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.568874 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krnvd\" (UniqueName: \"kubernetes.io/projected/60fc3ace-6591-4551-96ba-a746eacf5609-kube-api-access-krnvd\") pod \"60fc3ace-6591-4551-96ba-a746eacf5609\" (UID: \"60fc3ace-6591-4551-96ba-a746eacf5609\") " Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.568986 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lc8pr\" (UniqueName: \"kubernetes.io/projected/cc03cbc0-add4-4e90-a74a-edc369ee88b9-kube-api-access-lc8pr\") pod \"cc03cbc0-add4-4e90-a74a-edc369ee88b9\" (UID: \"cc03cbc0-add4-4e90-a74a-edc369ee88b9\") " Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.569058 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4j4s4\" (UniqueName: \"kubernetes.io/projected/a9ad29a8-c449-4ac3-a30d-64979050a1b0-kube-api-access-4j4s4\") pod \"a9ad29a8-c449-4ac3-a30d-64979050a1b0\" (UID: \"a9ad29a8-c449-4ac3-a30d-64979050a1b0\") " Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.569113 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/147cf615-2bd3-4db4-8d5a-40711c2b8f92-operator-scripts\") pod \"147cf615-2bd3-4db4-8d5a-40711c2b8f92\" (UID: \"147cf615-2bd3-4db4-8d5a-40711c2b8f92\") " Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.569161 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9ad29a8-c449-4ac3-a30d-64979050a1b0-operator-scripts\") pod \"a9ad29a8-c449-4ac3-a30d-64979050a1b0\" (UID: \"a9ad29a8-c449-4ac3-a30d-64979050a1b0\") " Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.569221 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc03cbc0-add4-4e90-a74a-edc369ee88b9-operator-scripts\") pod \"cc03cbc0-add4-4e90-a74a-edc369ee88b9\" (UID: \"cc03cbc0-add4-4e90-a74a-edc369ee88b9\") " Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.569246 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqsqj\" (UniqueName: \"kubernetes.io/projected/147cf615-2bd3-4db4-8d5a-40711c2b8f92-kube-api-access-pqsqj\") pod \"147cf615-2bd3-4db4-8d5a-40711c2b8f92\" (UID: \"147cf615-2bd3-4db4-8d5a-40711c2b8f92\") " Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.569290 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2v6l9\" (UniqueName: \"kubernetes.io/projected/d879cd8e-9523-47ac-acfc-75aa7fde0bc0-kube-api-access-2v6l9\") pod \"d879cd8e-9523-47ac-acfc-75aa7fde0bc0\" (UID: \"d879cd8e-9523-47ac-acfc-75aa7fde0bc0\") " Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.569324 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d879cd8e-9523-47ac-acfc-75aa7fde0bc0-operator-scripts\") pod \"d879cd8e-9523-47ac-acfc-75aa7fde0bc0\" (UID: \"d879cd8e-9523-47ac-acfc-75aa7fde0bc0\") " Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.569981 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/147cf615-2bd3-4db4-8d5a-40711c2b8f92-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "147cf615-2bd3-4db4-8d5a-40711c2b8f92" (UID: "147cf615-2bd3-4db4-8d5a-40711c2b8f92"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.570433 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d879cd8e-9523-47ac-acfc-75aa7fde0bc0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d879cd8e-9523-47ac-acfc-75aa7fde0bc0" (UID: "d879cd8e-9523-47ac-acfc-75aa7fde0bc0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.570922 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc03cbc0-add4-4e90-a74a-edc369ee88b9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cc03cbc0-add4-4e90-a74a-edc369ee88b9" (UID: "cc03cbc0-add4-4e90-a74a-edc369ee88b9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.571007 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9ad29a8-c449-4ac3-a30d-64979050a1b0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a9ad29a8-c449-4ac3-a30d-64979050a1b0" (UID: "a9ad29a8-c449-4ac3-a30d-64979050a1b0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.571482 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60fc3ace-6591-4551-96ba-a746eacf5609-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "60fc3ace-6591-4551-96ba-a746eacf5609" (UID: "60fc3ace-6591-4551-96ba-a746eacf5609"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.576423 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/147cf615-2bd3-4db4-8d5a-40711c2b8f92-kube-api-access-pqsqj" (OuterVolumeSpecName: "kube-api-access-pqsqj") pod "147cf615-2bd3-4db4-8d5a-40711c2b8f92" (UID: "147cf615-2bd3-4db4-8d5a-40711c2b8f92"). InnerVolumeSpecName "kube-api-access-pqsqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.576891 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc03cbc0-add4-4e90-a74a-edc369ee88b9-kube-api-access-lc8pr" (OuterVolumeSpecName: "kube-api-access-lc8pr") pod "cc03cbc0-add4-4e90-a74a-edc369ee88b9" (UID: "cc03cbc0-add4-4e90-a74a-edc369ee88b9"). InnerVolumeSpecName "kube-api-access-lc8pr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.577164 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60fc3ace-6591-4551-96ba-a746eacf5609-kube-api-access-krnvd" (OuterVolumeSpecName: "kube-api-access-krnvd") pod "60fc3ace-6591-4551-96ba-a746eacf5609" (UID: "60fc3ace-6591-4551-96ba-a746eacf5609"). InnerVolumeSpecName "kube-api-access-krnvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.577615 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9ad29a8-c449-4ac3-a30d-64979050a1b0-kube-api-access-4j4s4" (OuterVolumeSpecName: "kube-api-access-4j4s4") pod "a9ad29a8-c449-4ac3-a30d-64979050a1b0" (UID: "a9ad29a8-c449-4ac3-a30d-64979050a1b0"). InnerVolumeSpecName "kube-api-access-4j4s4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.579027 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d879cd8e-9523-47ac-acfc-75aa7fde0bc0-kube-api-access-2v6l9" (OuterVolumeSpecName: "kube-api-access-2v6l9") pod "d879cd8e-9523-47ac-acfc-75aa7fde0bc0" (UID: "d879cd8e-9523-47ac-acfc-75aa7fde0bc0"). InnerVolumeSpecName "kube-api-access-2v6l9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.670338 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86b3ecec-fc53-4d2d-843b-ab24ea43d82a-operator-scripts\") pod \"86b3ecec-fc53-4d2d-843b-ab24ea43d82a\" (UID: \"86b3ecec-fc53-4d2d-843b-ab24ea43d82a\") " Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.670669 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fw6cm\" (UniqueName: \"kubernetes.io/projected/86b3ecec-fc53-4d2d-843b-ab24ea43d82a-kube-api-access-fw6cm\") pod \"86b3ecec-fc53-4d2d-843b-ab24ea43d82a\" (UID: \"86b3ecec-fc53-4d2d-843b-ab24ea43d82a\") " Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.671002 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d879cd8e-9523-47ac-acfc-75aa7fde0bc0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.671020 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60fc3ace-6591-4551-96ba-a746eacf5609-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.671029 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krnvd\" (UniqueName: \"kubernetes.io/projected/60fc3ace-6591-4551-96ba-a746eacf5609-kube-api-access-krnvd\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.671041 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lc8pr\" (UniqueName: \"kubernetes.io/projected/cc03cbc0-add4-4e90-a74a-edc369ee88b9-kube-api-access-lc8pr\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.671051 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4j4s4\" (UniqueName: \"kubernetes.io/projected/a9ad29a8-c449-4ac3-a30d-64979050a1b0-kube-api-access-4j4s4\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.671061 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/147cf615-2bd3-4db4-8d5a-40711c2b8f92-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.671069 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9ad29a8-c449-4ac3-a30d-64979050a1b0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.671078 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqsqj\" (UniqueName: \"kubernetes.io/projected/147cf615-2bd3-4db4-8d5a-40711c2b8f92-kube-api-access-pqsqj\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.671086 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc03cbc0-add4-4e90-a74a-edc369ee88b9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.671094 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2v6l9\" (UniqueName: \"kubernetes.io/projected/d879cd8e-9523-47ac-acfc-75aa7fde0bc0-kube-api-access-2v6l9\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.671700 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86b3ecec-fc53-4d2d-843b-ab24ea43d82a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "86b3ecec-fc53-4d2d-843b-ab24ea43d82a" (UID: "86b3ecec-fc53-4d2d-843b-ab24ea43d82a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.673642 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86b3ecec-fc53-4d2d-843b-ab24ea43d82a-kube-api-access-fw6cm" (OuterVolumeSpecName: "kube-api-access-fw6cm") pod "86b3ecec-fc53-4d2d-843b-ab24ea43d82a" (UID: "86b3ecec-fc53-4d2d-843b-ab24ea43d82a"). InnerVolumeSpecName "kube-api-access-fw6cm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.772281 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86b3ecec-fc53-4d2d-843b-ab24ea43d82a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:34 crc kubenswrapper[4818]: I1122 05:10:34.772312 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fw6cm\" (UniqueName: \"kubernetes.io/projected/86b3ecec-fc53-4d2d-843b-ab24ea43d82a-kube-api-access-fw6cm\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:35 crc kubenswrapper[4818]: I1122 05:10:35.567366 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xdpgh" event={"ID":"06f95278-ba08-4ccd-af26-48c0c06157e7","Type":"ContainerStarted","Data":"ce5d3f0ff50a531c3adbc951f081ba274f9c58f5332864a1dba917c7fa062321"} Nov 22 05:10:35 crc kubenswrapper[4818]: I1122 05:10:35.574797 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bacb-account-create-grbqp" Nov 22 05:10:35 crc kubenswrapper[4818]: I1122 05:10:35.574784 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jpkvr" event={"ID":"486177c7-b0c3-45ba-a18c-79509154c9df","Type":"ContainerStarted","Data":"566e1df2e1e8ff694af911b80125bc51fa9ebcaae0be783202453e0a0427e77f"} Nov 22 05:10:35 crc kubenswrapper[4818]: I1122 05:10:35.600445 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-xdpgh" podStartSLOduration=2.191291473 podStartE2EDuration="8.600419993s" podCreationTimestamp="2025-11-22 05:10:27 +0000 UTC" firstStartedPulling="2025-11-22 05:10:28.179235647 +0000 UTC m=+1380.753652174" lastFinishedPulling="2025-11-22 05:10:34.588364167 +0000 UTC m=+1387.162780694" observedRunningTime="2025-11-22 05:10:35.588520057 +0000 UTC m=+1388.162936584" watchObservedRunningTime="2025-11-22 05:10:35.600419993 +0000 UTC m=+1388.174836530" Nov 22 05:10:35 crc kubenswrapper[4818]: I1122 05:10:35.611007 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-jpkvr" podStartSLOduration=2.912180188 podStartE2EDuration="40.610985734s" podCreationTimestamp="2025-11-22 05:09:55 +0000 UTC" firstStartedPulling="2025-11-22 05:09:56.889333135 +0000 UTC m=+1349.463749662" lastFinishedPulling="2025-11-22 05:10:34.588138661 +0000 UTC m=+1387.162555208" observedRunningTime="2025-11-22 05:10:35.603425983 +0000 UTC m=+1388.177842510" watchObservedRunningTime="2025-11-22 05:10:35.610985734 +0000 UTC m=+1388.185402281" Nov 22 05:10:40 crc kubenswrapper[4818]: I1122 05:10:40.613108 4818 generic.go:334] "Generic (PLEG): container finished" podID="06f95278-ba08-4ccd-af26-48c0c06157e7" containerID="ce5d3f0ff50a531c3adbc951f081ba274f9c58f5332864a1dba917c7fa062321" exitCode=0 Nov 22 05:10:40 crc kubenswrapper[4818]: I1122 05:10:40.613225 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xdpgh" event={"ID":"06f95278-ba08-4ccd-af26-48c0c06157e7","Type":"ContainerDied","Data":"ce5d3f0ff50a531c3adbc951f081ba274f9c58f5332864a1dba917c7fa062321"} Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.087553 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xdpgh" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.207281 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06f95278-ba08-4ccd-af26-48c0c06157e7-config-data\") pod \"06f95278-ba08-4ccd-af26-48c0c06157e7\" (UID: \"06f95278-ba08-4ccd-af26-48c0c06157e7\") " Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.207485 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f95278-ba08-4ccd-af26-48c0c06157e7-combined-ca-bundle\") pod \"06f95278-ba08-4ccd-af26-48c0c06157e7\" (UID: \"06f95278-ba08-4ccd-af26-48c0c06157e7\") " Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.207572 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpzg6\" (UniqueName: \"kubernetes.io/projected/06f95278-ba08-4ccd-af26-48c0c06157e7-kube-api-access-bpzg6\") pod \"06f95278-ba08-4ccd-af26-48c0c06157e7\" (UID: \"06f95278-ba08-4ccd-af26-48c0c06157e7\") " Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.222713 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06f95278-ba08-4ccd-af26-48c0c06157e7-kube-api-access-bpzg6" (OuterVolumeSpecName: "kube-api-access-bpzg6") pod "06f95278-ba08-4ccd-af26-48c0c06157e7" (UID: "06f95278-ba08-4ccd-af26-48c0c06157e7"). InnerVolumeSpecName "kube-api-access-bpzg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.231848 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f95278-ba08-4ccd-af26-48c0c06157e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06f95278-ba08-4ccd-af26-48c0c06157e7" (UID: "06f95278-ba08-4ccd-af26-48c0c06157e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.253384 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f95278-ba08-4ccd-af26-48c0c06157e7-config-data" (OuterVolumeSpecName: "config-data") pod "06f95278-ba08-4ccd-af26-48c0c06157e7" (UID: "06f95278-ba08-4ccd-af26-48c0c06157e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.309344 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06f95278-ba08-4ccd-af26-48c0c06157e7-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.309373 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f95278-ba08-4ccd-af26-48c0c06157e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.309385 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpzg6\" (UniqueName: \"kubernetes.io/projected/06f95278-ba08-4ccd-af26-48c0c06157e7-kube-api-access-bpzg6\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.632090 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xdpgh" event={"ID":"06f95278-ba08-4ccd-af26-48c0c06157e7","Type":"ContainerDied","Data":"3d8a02ed5bd25dfbcb523c84dd32956c77684be759aab7ca584d53c55e3a8c12"} Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.632130 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d8a02ed5bd25dfbcb523c84dd32956c77684be759aab7ca584d53c55e3a8c12" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.632181 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xdpgh" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.886490 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-t5p4f"] Nov 22 05:10:42 crc kubenswrapper[4818]: E1122 05:10:42.886852 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60fc3ace-6591-4551-96ba-a746eacf5609" containerName="mariadb-database-create" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.886866 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="60fc3ace-6591-4551-96ba-a746eacf5609" containerName="mariadb-database-create" Nov 22 05:10:42 crc kubenswrapper[4818]: E1122 05:10:42.886875 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d879cd8e-9523-47ac-acfc-75aa7fde0bc0" containerName="mariadb-account-create" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.886881 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="d879cd8e-9523-47ac-acfc-75aa7fde0bc0" containerName="mariadb-account-create" Nov 22 05:10:42 crc kubenswrapper[4818]: E1122 05:10:42.886889 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86b3ecec-fc53-4d2d-843b-ab24ea43d82a" containerName="mariadb-account-create" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.886895 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="86b3ecec-fc53-4d2d-843b-ab24ea43d82a" containerName="mariadb-account-create" Nov 22 05:10:42 crc kubenswrapper[4818]: E1122 05:10:42.886904 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f95278-ba08-4ccd-af26-48c0c06157e7" containerName="keystone-db-sync" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.886911 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f95278-ba08-4ccd-af26-48c0c06157e7" containerName="keystone-db-sync" Nov 22 05:10:42 crc kubenswrapper[4818]: E1122 05:10:42.886922 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ad29a8-c449-4ac3-a30d-64979050a1b0" containerName="mariadb-database-create" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.886928 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ad29a8-c449-4ac3-a30d-64979050a1b0" containerName="mariadb-database-create" Nov 22 05:10:42 crc kubenswrapper[4818]: E1122 05:10:42.886938 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="147cf615-2bd3-4db4-8d5a-40711c2b8f92" containerName="mariadb-database-create" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.886943 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="147cf615-2bd3-4db4-8d5a-40711c2b8f92" containerName="mariadb-database-create" Nov 22 05:10:42 crc kubenswrapper[4818]: E1122 05:10:42.886959 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc03cbc0-add4-4e90-a74a-edc369ee88b9" containerName="mariadb-account-create" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.886965 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc03cbc0-add4-4e90-a74a-edc369ee88b9" containerName="mariadb-account-create" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.887105 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="d879cd8e-9523-47ac-acfc-75aa7fde0bc0" containerName="mariadb-account-create" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.887124 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="147cf615-2bd3-4db4-8d5a-40711c2b8f92" containerName="mariadb-database-create" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.887139 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc03cbc0-add4-4e90-a74a-edc369ee88b9" containerName="mariadb-account-create" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.887148 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="60fc3ace-6591-4551-96ba-a746eacf5609" containerName="mariadb-database-create" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.887156 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="06f95278-ba08-4ccd-af26-48c0c06157e7" containerName="keystone-db-sync" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.887164 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="86b3ecec-fc53-4d2d-843b-ab24ea43d82a" containerName="mariadb-account-create" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.887173 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9ad29a8-c449-4ac3-a30d-64979050a1b0" containerName="mariadb-database-create" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.887679 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.890406 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.890678 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.890856 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-tds7x" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.890967 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.892575 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.914043 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-t5p4f"] Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.923598 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75bb4695fc-zpn6h"] Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.924962 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" Nov 22 05:10:42 crc kubenswrapper[4818]: I1122 05:10:42.940736 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75bb4695fc-zpn6h"] Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.021900 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-combined-ca-bundle\") pod \"keystone-bootstrap-t5p4f\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.022001 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-credential-keys\") pod \"keystone-bootstrap-t5p4f\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.022040 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-dns-svc\") pod \"dnsmasq-dns-75bb4695fc-zpn6h\" (UID: \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\") " pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.022077 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-config-data\") pod \"keystone-bootstrap-t5p4f\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.022128 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t56rx\" (UniqueName: \"kubernetes.io/projected/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-kube-api-access-t56rx\") pod \"dnsmasq-dns-75bb4695fc-zpn6h\" (UID: \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\") " pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.022180 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-fernet-keys\") pod \"keystone-bootstrap-t5p4f\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.022206 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-config\") pod \"dnsmasq-dns-75bb4695fc-zpn6h\" (UID: \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\") " pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.022290 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92hgm\" (UniqueName: \"kubernetes.io/projected/54ceda18-1bdf-4adf-bb72-3c323f693eb0-kube-api-access-92hgm\") pod \"keystone-bootstrap-t5p4f\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.022326 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-ovsdbserver-nb\") pod \"dnsmasq-dns-75bb4695fc-zpn6h\" (UID: \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\") " pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.022385 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-ovsdbserver-sb\") pod \"dnsmasq-dns-75bb4695fc-zpn6h\" (UID: \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\") " pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.022415 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-scripts\") pod \"keystone-bootstrap-t5p4f\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.126599 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-combined-ca-bundle\") pod \"keystone-bootstrap-t5p4f\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.127354 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-credential-keys\") pod \"keystone-bootstrap-t5p4f\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.127416 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-dns-svc\") pod \"dnsmasq-dns-75bb4695fc-zpn6h\" (UID: \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\") " pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.127467 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-config-data\") pod \"keystone-bootstrap-t5p4f\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.127511 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t56rx\" (UniqueName: \"kubernetes.io/projected/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-kube-api-access-t56rx\") pod \"dnsmasq-dns-75bb4695fc-zpn6h\" (UID: \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\") " pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.127582 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-fernet-keys\") pod \"keystone-bootstrap-t5p4f\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.127616 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-config\") pod \"dnsmasq-dns-75bb4695fc-zpn6h\" (UID: \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\") " pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.127655 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92hgm\" (UniqueName: \"kubernetes.io/projected/54ceda18-1bdf-4adf-bb72-3c323f693eb0-kube-api-access-92hgm\") pod \"keystone-bootstrap-t5p4f\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.127688 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-ovsdbserver-nb\") pod \"dnsmasq-dns-75bb4695fc-zpn6h\" (UID: \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\") " pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.127742 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-ovsdbserver-sb\") pod \"dnsmasq-dns-75bb4695fc-zpn6h\" (UID: \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\") " pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.127781 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-scripts\") pod \"keystone-bootstrap-t5p4f\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.129914 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-7spb7"] Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.132960 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7spb7" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.134211 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-config\") pod \"dnsmasq-dns-75bb4695fc-zpn6h\" (UID: \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\") " pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.134843 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-fernet-keys\") pod \"keystone-bootstrap-t5p4f\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.134892 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-ovsdbserver-nb\") pod \"dnsmasq-dns-75bb4695fc-zpn6h\" (UID: \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\") " pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.137368 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-dns-svc\") pod \"dnsmasq-dns-75bb4695fc-zpn6h\" (UID: \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\") " pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.140933 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-ovsdbserver-sb\") pod \"dnsmasq-dns-75bb4695fc-zpn6h\" (UID: \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\") " pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.165940 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-credential-keys\") pod \"keystone-bootstrap-t5p4f\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.166102 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.166182 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-9pzqx" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.167009 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-scripts\") pod \"keystone-bootstrap-t5p4f\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.168472 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.185174 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7spb7"] Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.190392 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-combined-ca-bundle\") pod \"keystone-bootstrap-t5p4f\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.190427 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-config-data\") pod \"keystone-bootstrap-t5p4f\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.197991 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92hgm\" (UniqueName: \"kubernetes.io/projected/54ceda18-1bdf-4adf-bb72-3c323f693eb0-kube-api-access-92hgm\") pod \"keystone-bootstrap-t5p4f\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.198627 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t56rx\" (UniqueName: \"kubernetes.io/projected/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-kube-api-access-t56rx\") pod \"dnsmasq-dns-75bb4695fc-zpn6h\" (UID: \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\") " pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.210465 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.244912 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.247306 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.248365 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwvrh\" (UniqueName: \"kubernetes.io/projected/0ca1db80-7d8d-44cf-80bf-afa1c0798182-kube-api-access-nwvrh\") pod \"cinder-db-sync-7spb7\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " pod="openstack/cinder-db-sync-7spb7" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.248393 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-config-data\") pod \"cinder-db-sync-7spb7\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " pod="openstack/cinder-db-sync-7spb7" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.248418 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-db-sync-config-data\") pod \"cinder-db-sync-7spb7\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " pod="openstack/cinder-db-sync-7spb7" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.248439 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-scripts\") pod \"cinder-db-sync-7spb7\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " pod="openstack/cinder-db-sync-7spb7" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.248499 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0ca1db80-7d8d-44cf-80bf-afa1c0798182-etc-machine-id\") pod \"cinder-db-sync-7spb7\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " pod="openstack/cinder-db-sync-7spb7" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.248523 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-combined-ca-bundle\") pod \"cinder-db-sync-7spb7\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " pod="openstack/cinder-db-sync-7spb7" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.253639 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.256240 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.256472 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.280624 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-rd5pw"] Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.281732 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-rd5pw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.288745 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.289053 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.290651 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-jn2fl" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.330348 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-rd5pw"] Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.342440 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.349842 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.349921 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0ca1db80-7d8d-44cf-80bf-afa1c0798182-etc-machine-id\") pod \"cinder-db-sync-7spb7\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " pod="openstack/cinder-db-sync-7spb7" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.349973 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-combined-ca-bundle\") pod \"cinder-db-sync-7spb7\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " pod="openstack/cinder-db-sync-7spb7" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.350009 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f84522a7-f356-40c9-8ea0-13b1503993fe-combined-ca-bundle\") pod \"neutron-db-sync-rd5pw\" (UID: \"f84522a7-f356-40c9-8ea0-13b1503993fe\") " pod="openstack/neutron-db-sync-rd5pw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.350041 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-scripts\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.350090 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.350112 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x826k\" (UniqueName: \"kubernetes.io/projected/0f6203d4-9fff-483c-bc5c-84deb159d47d-kube-api-access-x826k\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.350144 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwvrh\" (UniqueName: \"kubernetes.io/projected/0ca1db80-7d8d-44cf-80bf-afa1c0798182-kube-api-access-nwvrh\") pod \"cinder-db-sync-7spb7\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " pod="openstack/cinder-db-sync-7spb7" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.350169 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-config-data\") pod \"cinder-db-sync-7spb7\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " pod="openstack/cinder-db-sync-7spb7" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.350192 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njwqs\" (UniqueName: \"kubernetes.io/projected/f84522a7-f356-40c9-8ea0-13b1503993fe-kube-api-access-njwqs\") pod \"neutron-db-sync-rd5pw\" (UID: \"f84522a7-f356-40c9-8ea0-13b1503993fe\") " pod="openstack/neutron-db-sync-rd5pw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.350240 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-db-sync-config-data\") pod \"cinder-db-sync-7spb7\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " pod="openstack/cinder-db-sync-7spb7" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.350278 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f6203d4-9fff-483c-bc5c-84deb159d47d-log-httpd\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.350307 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-scripts\") pod \"cinder-db-sync-7spb7\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " pod="openstack/cinder-db-sync-7spb7" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.350334 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f6203d4-9fff-483c-bc5c-84deb159d47d-run-httpd\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.350376 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-config-data\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.350400 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f84522a7-f356-40c9-8ea0-13b1503993fe-config\") pod \"neutron-db-sync-rd5pw\" (UID: \"f84522a7-f356-40c9-8ea0-13b1503993fe\") " pod="openstack/neutron-db-sync-rd5pw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.350504 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0ca1db80-7d8d-44cf-80bf-afa1c0798182-etc-machine-id\") pod \"cinder-db-sync-7spb7\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " pod="openstack/cinder-db-sync-7spb7" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.365225 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-n2ldz"] Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.365667 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-db-sync-config-data\") pod \"cinder-db-sync-7spb7\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " pod="openstack/cinder-db-sync-7spb7" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.366417 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-n2ldz" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.370789 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.371483 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-j9jxn" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.372749 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75bb4695fc-zpn6h"] Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.372836 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-config-data\") pod \"cinder-db-sync-7spb7\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " pod="openstack/cinder-db-sync-7spb7" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.382297 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwvrh\" (UniqueName: \"kubernetes.io/projected/0ca1db80-7d8d-44cf-80bf-afa1c0798182-kube-api-access-nwvrh\") pod \"cinder-db-sync-7spb7\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " pod="openstack/cinder-db-sync-7spb7" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.385732 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-combined-ca-bundle\") pod \"cinder-db-sync-7spb7\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " pod="openstack/cinder-db-sync-7spb7" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.387741 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-scripts\") pod \"cinder-db-sync-7spb7\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " pod="openstack/cinder-db-sync-7spb7" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.403514 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-n2ldz"] Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.414375 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-qnwqb"] Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.416088 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qnwqb" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.427294 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-qnwqb"] Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.427447 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-nnwmw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.428505 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.428717 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.432144 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-745b9ddc8c-cpmmw"] Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.433802 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.456294 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.456334 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x826k\" (UniqueName: \"kubernetes.io/projected/0f6203d4-9fff-483c-bc5c-84deb159d47d-kube-api-access-x826k\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.456375 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njwqs\" (UniqueName: \"kubernetes.io/projected/f84522a7-f356-40c9-8ea0-13b1503993fe-kube-api-access-njwqs\") pod \"neutron-db-sync-rd5pw\" (UID: \"f84522a7-f356-40c9-8ea0-13b1503993fe\") " pod="openstack/neutron-db-sync-rd5pw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.456403 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f6203d4-9fff-483c-bc5c-84deb159d47d-log-httpd\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.456432 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f6203d4-9fff-483c-bc5c-84deb159d47d-run-httpd\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.456472 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-config-data\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.456495 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f84522a7-f356-40c9-8ea0-13b1503993fe-config\") pod \"neutron-db-sync-rd5pw\" (UID: \"f84522a7-f356-40c9-8ea0-13b1503993fe\") " pod="openstack/neutron-db-sync-rd5pw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.456530 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.456604 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f84522a7-f356-40c9-8ea0-13b1503993fe-combined-ca-bundle\") pod \"neutron-db-sync-rd5pw\" (UID: \"f84522a7-f356-40c9-8ea0-13b1503993fe\") " pod="openstack/neutron-db-sync-rd5pw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.456634 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-scripts\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.457695 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f6203d4-9fff-483c-bc5c-84deb159d47d-log-httpd\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.458704 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f6203d4-9fff-483c-bc5c-84deb159d47d-run-httpd\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.464444 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-scripts\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.467850 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f84522a7-f356-40c9-8ea0-13b1503993fe-combined-ca-bundle\") pod \"neutron-db-sync-rd5pw\" (UID: \"f84522a7-f356-40c9-8ea0-13b1503993fe\") " pod="openstack/neutron-db-sync-rd5pw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.468048 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.470692 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.473098 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-config-data\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.486623 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-745b9ddc8c-cpmmw"] Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.489472 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njwqs\" (UniqueName: \"kubernetes.io/projected/f84522a7-f356-40c9-8ea0-13b1503993fe-kube-api-access-njwqs\") pod \"neutron-db-sync-rd5pw\" (UID: \"f84522a7-f356-40c9-8ea0-13b1503993fe\") " pod="openstack/neutron-db-sync-rd5pw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.490228 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x826k\" (UniqueName: \"kubernetes.io/projected/0f6203d4-9fff-483c-bc5c-84deb159d47d-kube-api-access-x826k\") pod \"ceilometer-0\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.495319 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f84522a7-f356-40c9-8ea0-13b1503993fe-config\") pod \"neutron-db-sync-rd5pw\" (UID: \"f84522a7-f356-40c9-8ea0-13b1503993fe\") " pod="openstack/neutron-db-sync-rd5pw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.558342 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzf5g\" (UniqueName: \"kubernetes.io/projected/e86e2277-a801-4562-8741-84420e0d8b5e-kube-api-access-gzf5g\") pod \"placement-db-sync-qnwqb\" (UID: \"e86e2277-a801-4562-8741-84420e0d8b5e\") " pod="openstack/placement-db-sync-qnwqb" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.558400 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-config\") pod \"dnsmasq-dns-745b9ddc8c-cpmmw\" (UID: \"828b46a2-0bd7-430c-8cd1-06be3d81b963\") " pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.558421 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-dns-svc\") pod \"dnsmasq-dns-745b9ddc8c-cpmmw\" (UID: \"828b46a2-0bd7-430c-8cd1-06be3d81b963\") " pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.558445 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a8a8454-074a-48c1-b1f2-f800dd6f0f0a-combined-ca-bundle\") pod \"barbican-db-sync-n2ldz\" (UID: \"2a8a8454-074a-48c1-b1f2-f800dd6f0f0a\") " pod="openstack/barbican-db-sync-n2ldz" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.558467 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-577sx\" (UniqueName: \"kubernetes.io/projected/2a8a8454-074a-48c1-b1f2-f800dd6f0f0a-kube-api-access-577sx\") pod \"barbican-db-sync-n2ldz\" (UID: \"2a8a8454-074a-48c1-b1f2-f800dd6f0f0a\") " pod="openstack/barbican-db-sync-n2ldz" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.558485 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-ovsdbserver-nb\") pod \"dnsmasq-dns-745b9ddc8c-cpmmw\" (UID: \"828b46a2-0bd7-430c-8cd1-06be3d81b963\") " pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.558514 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxcvc\" (UniqueName: \"kubernetes.io/projected/828b46a2-0bd7-430c-8cd1-06be3d81b963-kube-api-access-rxcvc\") pod \"dnsmasq-dns-745b9ddc8c-cpmmw\" (UID: \"828b46a2-0bd7-430c-8cd1-06be3d81b963\") " pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.558566 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86e2277-a801-4562-8741-84420e0d8b5e-combined-ca-bundle\") pod \"placement-db-sync-qnwqb\" (UID: \"e86e2277-a801-4562-8741-84420e0d8b5e\") " pod="openstack/placement-db-sync-qnwqb" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.558597 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2a8a8454-074a-48c1-b1f2-f800dd6f0f0a-db-sync-config-data\") pod \"barbican-db-sync-n2ldz\" (UID: \"2a8a8454-074a-48c1-b1f2-f800dd6f0f0a\") " pod="openstack/barbican-db-sync-n2ldz" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.558629 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e86e2277-a801-4562-8741-84420e0d8b5e-config-data\") pod \"placement-db-sync-qnwqb\" (UID: \"e86e2277-a801-4562-8741-84420e0d8b5e\") " pod="openstack/placement-db-sync-qnwqb" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.558647 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e86e2277-a801-4562-8741-84420e0d8b5e-scripts\") pod \"placement-db-sync-qnwqb\" (UID: \"e86e2277-a801-4562-8741-84420e0d8b5e\") " pod="openstack/placement-db-sync-qnwqb" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.558666 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-ovsdbserver-sb\") pod \"dnsmasq-dns-745b9ddc8c-cpmmw\" (UID: \"828b46a2-0bd7-430c-8cd1-06be3d81b963\") " pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.558684 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e86e2277-a801-4562-8741-84420e0d8b5e-logs\") pod \"placement-db-sync-qnwqb\" (UID: \"e86e2277-a801-4562-8741-84420e0d8b5e\") " pod="openstack/placement-db-sync-qnwqb" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.569446 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7spb7" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.604747 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.660136 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxcvc\" (UniqueName: \"kubernetes.io/projected/828b46a2-0bd7-430c-8cd1-06be3d81b963-kube-api-access-rxcvc\") pod \"dnsmasq-dns-745b9ddc8c-cpmmw\" (UID: \"828b46a2-0bd7-430c-8cd1-06be3d81b963\") " pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.660206 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86e2277-a801-4562-8741-84420e0d8b5e-combined-ca-bundle\") pod \"placement-db-sync-qnwqb\" (UID: \"e86e2277-a801-4562-8741-84420e0d8b5e\") " pod="openstack/placement-db-sync-qnwqb" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.660232 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2a8a8454-074a-48c1-b1f2-f800dd6f0f0a-db-sync-config-data\") pod \"barbican-db-sync-n2ldz\" (UID: \"2a8a8454-074a-48c1-b1f2-f800dd6f0f0a\") " pod="openstack/barbican-db-sync-n2ldz" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.660275 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e86e2277-a801-4562-8741-84420e0d8b5e-config-data\") pod \"placement-db-sync-qnwqb\" (UID: \"e86e2277-a801-4562-8741-84420e0d8b5e\") " pod="openstack/placement-db-sync-qnwqb" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.660290 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e86e2277-a801-4562-8741-84420e0d8b5e-scripts\") pod \"placement-db-sync-qnwqb\" (UID: \"e86e2277-a801-4562-8741-84420e0d8b5e\") " pod="openstack/placement-db-sync-qnwqb" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.660306 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-ovsdbserver-sb\") pod \"dnsmasq-dns-745b9ddc8c-cpmmw\" (UID: \"828b46a2-0bd7-430c-8cd1-06be3d81b963\") " pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.660327 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e86e2277-a801-4562-8741-84420e0d8b5e-logs\") pod \"placement-db-sync-qnwqb\" (UID: \"e86e2277-a801-4562-8741-84420e0d8b5e\") " pod="openstack/placement-db-sync-qnwqb" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.660360 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzf5g\" (UniqueName: \"kubernetes.io/projected/e86e2277-a801-4562-8741-84420e0d8b5e-kube-api-access-gzf5g\") pod \"placement-db-sync-qnwqb\" (UID: \"e86e2277-a801-4562-8741-84420e0d8b5e\") " pod="openstack/placement-db-sync-qnwqb" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.660391 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-config\") pod \"dnsmasq-dns-745b9ddc8c-cpmmw\" (UID: \"828b46a2-0bd7-430c-8cd1-06be3d81b963\") " pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.660409 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-dns-svc\") pod \"dnsmasq-dns-745b9ddc8c-cpmmw\" (UID: \"828b46a2-0bd7-430c-8cd1-06be3d81b963\") " pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.660431 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a8a8454-074a-48c1-b1f2-f800dd6f0f0a-combined-ca-bundle\") pod \"barbican-db-sync-n2ldz\" (UID: \"2a8a8454-074a-48c1-b1f2-f800dd6f0f0a\") " pod="openstack/barbican-db-sync-n2ldz" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.660452 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-577sx\" (UniqueName: \"kubernetes.io/projected/2a8a8454-074a-48c1-b1f2-f800dd6f0f0a-kube-api-access-577sx\") pod \"barbican-db-sync-n2ldz\" (UID: \"2a8a8454-074a-48c1-b1f2-f800dd6f0f0a\") " pod="openstack/barbican-db-sync-n2ldz" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.660472 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-ovsdbserver-nb\") pod \"dnsmasq-dns-745b9ddc8c-cpmmw\" (UID: \"828b46a2-0bd7-430c-8cd1-06be3d81b963\") " pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.661417 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-ovsdbserver-nb\") pod \"dnsmasq-dns-745b9ddc8c-cpmmw\" (UID: \"828b46a2-0bd7-430c-8cd1-06be3d81b963\") " pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.661570 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-ovsdbserver-sb\") pod \"dnsmasq-dns-745b9ddc8c-cpmmw\" (UID: \"828b46a2-0bd7-430c-8cd1-06be3d81b963\") " pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.661678 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e86e2277-a801-4562-8741-84420e0d8b5e-logs\") pod \"placement-db-sync-qnwqb\" (UID: \"e86e2277-a801-4562-8741-84420e0d8b5e\") " pod="openstack/placement-db-sync-qnwqb" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.663934 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-dns-svc\") pod \"dnsmasq-dns-745b9ddc8c-cpmmw\" (UID: \"828b46a2-0bd7-430c-8cd1-06be3d81b963\") " pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.664632 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-config\") pod \"dnsmasq-dns-745b9ddc8c-cpmmw\" (UID: \"828b46a2-0bd7-430c-8cd1-06be3d81b963\") " pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.670674 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2a8a8454-074a-48c1-b1f2-f800dd6f0f0a-db-sync-config-data\") pod \"barbican-db-sync-n2ldz\" (UID: \"2a8a8454-074a-48c1-b1f2-f800dd6f0f0a\") " pod="openstack/barbican-db-sync-n2ldz" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.671471 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86e2277-a801-4562-8741-84420e0d8b5e-combined-ca-bundle\") pod \"placement-db-sync-qnwqb\" (UID: \"e86e2277-a801-4562-8741-84420e0d8b5e\") " pod="openstack/placement-db-sync-qnwqb" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.671568 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e86e2277-a801-4562-8741-84420e0d8b5e-scripts\") pod \"placement-db-sync-qnwqb\" (UID: \"e86e2277-a801-4562-8741-84420e0d8b5e\") " pod="openstack/placement-db-sync-qnwqb" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.671634 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a8a8454-074a-48c1-b1f2-f800dd6f0f0a-combined-ca-bundle\") pod \"barbican-db-sync-n2ldz\" (UID: \"2a8a8454-074a-48c1-b1f2-f800dd6f0f0a\") " pod="openstack/barbican-db-sync-n2ldz" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.678267 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e86e2277-a801-4562-8741-84420e0d8b5e-config-data\") pod \"placement-db-sync-qnwqb\" (UID: \"e86e2277-a801-4562-8741-84420e0d8b5e\") " pod="openstack/placement-db-sync-qnwqb" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.687278 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-577sx\" (UniqueName: \"kubernetes.io/projected/2a8a8454-074a-48c1-b1f2-f800dd6f0f0a-kube-api-access-577sx\") pod \"barbican-db-sync-n2ldz\" (UID: \"2a8a8454-074a-48c1-b1f2-f800dd6f0f0a\") " pod="openstack/barbican-db-sync-n2ldz" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.690780 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzf5g\" (UniqueName: \"kubernetes.io/projected/e86e2277-a801-4562-8741-84420e0d8b5e-kube-api-access-gzf5g\") pod \"placement-db-sync-qnwqb\" (UID: \"e86e2277-a801-4562-8741-84420e0d8b5e\") " pod="openstack/placement-db-sync-qnwqb" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.694240 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxcvc\" (UniqueName: \"kubernetes.io/projected/828b46a2-0bd7-430c-8cd1-06be3d81b963-kube-api-access-rxcvc\") pod \"dnsmasq-dns-745b9ddc8c-cpmmw\" (UID: \"828b46a2-0bd7-430c-8cd1-06be3d81b963\") " pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.762540 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-rd5pw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.791799 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-n2ldz" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.807560 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qnwqb" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.822608 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:10:43 crc kubenswrapper[4818]: I1122 05:10:43.856817 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75bb4695fc-zpn6h"] Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.045312 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-t5p4f"] Nov 22 05:10:44 crc kubenswrapper[4818]: W1122 05:10:44.119313 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54ceda18_1bdf_4adf_bb72_3c323f693eb0.slice/crio-88878bee1c87320de1d8d989b85291174dbbeecc3c30dafb3206e1c2dd0d45b5 WatchSource:0}: Error finding container 88878bee1c87320de1d8d989b85291174dbbeecc3c30dafb3206e1c2dd0d45b5: Status 404 returned error can't find the container with id 88878bee1c87320de1d8d989b85291174dbbeecc3c30dafb3206e1c2dd0d45b5 Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.223276 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7spb7"] Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.238847 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.566509 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-rd5pw"] Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.574055 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-n2ldz"] Nov 22 05:10:44 crc kubenswrapper[4818]: W1122 05:10:44.576462 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a8a8454_074a_48c1_b1f2_f800dd6f0f0a.slice/crio-2874f707c5ab53823774fd6a42fca4daf3e1b8cd08ce7f7d50f19bdfe3ddf2ec WatchSource:0}: Error finding container 2874f707c5ab53823774fd6a42fca4daf3e1b8cd08ce7f7d50f19bdfe3ddf2ec: Status 404 returned error can't find the container with id 2874f707c5ab53823774fd6a42fca4daf3e1b8cd08ce7f7d50f19bdfe3ddf2ec Nov 22 05:10:44 crc kubenswrapper[4818]: W1122 05:10:44.580299 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode86e2277_a801_4562_8741_84420e0d8b5e.slice/crio-f0a234acc6ddcc6853ec1698a5a3e2b505688f4b92adc73f19d17e7e5605ca66 WatchSource:0}: Error finding container f0a234acc6ddcc6853ec1698a5a3e2b505688f4b92adc73f19d17e7e5605ca66: Status 404 returned error can't find the container with id f0a234acc6ddcc6853ec1698a5a3e2b505688f4b92adc73f19d17e7e5605ca66 Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.584829 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-qnwqb"] Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.655627 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7spb7" event={"ID":"0ca1db80-7d8d-44cf-80bf-afa1c0798182","Type":"ContainerStarted","Data":"a017ef0933330bf734fddc35ecf89b4f6e30c951af50a73dd41969f663162b68"} Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.657223 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qnwqb" event={"ID":"e86e2277-a801-4562-8741-84420e0d8b5e","Type":"ContainerStarted","Data":"f0a234acc6ddcc6853ec1698a5a3e2b505688f4b92adc73f19d17e7e5605ca66"} Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.658536 4818 generic.go:334] "Generic (PLEG): container finished" podID="daf1dc83-5f20-4f49-8327-8d0bd5e11bc0" containerID="745b8b5a96456ef61afb5180857d09a32abeac9d1af1974e16409d1abf69594e" exitCode=0 Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.658584 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" event={"ID":"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0","Type":"ContainerDied","Data":"745b8b5a96456ef61afb5180857d09a32abeac9d1af1974e16409d1abf69594e"} Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.658605 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" event={"ID":"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0","Type":"ContainerStarted","Data":"e044cf7043e751ff88f4d9241c3e7ef008a001acbb71993f9c933a77386fe1dd"} Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.660013 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-n2ldz" event={"ID":"2a8a8454-074a-48c1-b1f2-f800dd6f0f0a","Type":"ContainerStarted","Data":"2874f707c5ab53823774fd6a42fca4daf3e1b8cd08ce7f7d50f19bdfe3ddf2ec"} Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.661930 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t5p4f" event={"ID":"54ceda18-1bdf-4adf-bb72-3c323f693eb0","Type":"ContainerStarted","Data":"f7a4d0ea11175f50c5e59c18c7fb82d32bcd64c49e2e4b60395c59ecb1aa066d"} Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.661956 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t5p4f" event={"ID":"54ceda18-1bdf-4adf-bb72-3c323f693eb0","Type":"ContainerStarted","Data":"88878bee1c87320de1d8d989b85291174dbbeecc3c30dafb3206e1c2dd0d45b5"} Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.671157 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f6203d4-9fff-483c-bc5c-84deb159d47d","Type":"ContainerStarted","Data":"bfe8c2dc1c03792e20e9966cc8989dd047e089919fb0d47ee582099b9a40c8da"} Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.673021 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-rd5pw" event={"ID":"f84522a7-f356-40c9-8ea0-13b1503993fe","Type":"ContainerStarted","Data":"ba1eb2580d8e6f81a2ce1117a4ff385f4c956bd774860a8186f6bf80af7d14e7"} Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.715371 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-t5p4f" podStartSLOduration=2.715346053 podStartE2EDuration="2.715346053s" podCreationTimestamp="2025-11-22 05:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:10:44.699588403 +0000 UTC m=+1397.274004930" watchObservedRunningTime="2025-11-22 05:10:44.715346053 +0000 UTC m=+1397.289762570" Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.772373 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-745b9ddc8c-cpmmw"] Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.973955 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.997215 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t56rx\" (UniqueName: \"kubernetes.io/projected/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-kube-api-access-t56rx\") pod \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\" (UID: \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\") " Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.997419 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-dns-svc\") pod \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\" (UID: \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\") " Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.997527 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-ovsdbserver-sb\") pod \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\" (UID: \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\") " Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.997570 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-ovsdbserver-nb\") pod \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\" (UID: \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\") " Nov 22 05:10:44 crc kubenswrapper[4818]: I1122 05:10:44.997586 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-config\") pod \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\" (UID: \"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0\") " Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.004697 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-kube-api-access-t56rx" (OuterVolumeSpecName: "kube-api-access-t56rx") pod "daf1dc83-5f20-4f49-8327-8d0bd5e11bc0" (UID: "daf1dc83-5f20-4f49-8327-8d0bd5e11bc0"). InnerVolumeSpecName "kube-api-access-t56rx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.027265 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "daf1dc83-5f20-4f49-8327-8d0bd5e11bc0" (UID: "daf1dc83-5f20-4f49-8327-8d0bd5e11bc0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.030764 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "daf1dc83-5f20-4f49-8327-8d0bd5e11bc0" (UID: "daf1dc83-5f20-4f49-8327-8d0bd5e11bc0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.063119 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-config" (OuterVolumeSpecName: "config") pod "daf1dc83-5f20-4f49-8327-8d0bd5e11bc0" (UID: "daf1dc83-5f20-4f49-8327-8d0bd5e11bc0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.063487 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "daf1dc83-5f20-4f49-8327-8d0bd5e11bc0" (UID: "daf1dc83-5f20-4f49-8327-8d0bd5e11bc0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.065381 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.107579 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.107618 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.107633 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.107645 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.107656 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t56rx\" (UniqueName: \"kubernetes.io/projected/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0-kube-api-access-t56rx\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.686655 4818 generic.go:334] "Generic (PLEG): container finished" podID="828b46a2-0bd7-430c-8cd1-06be3d81b963" containerID="c56db2f1c6f7d67c272a1235b8914e89669824b3796414f355ec5eb8bad20cfd" exitCode=0 Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.686811 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" event={"ID":"828b46a2-0bd7-430c-8cd1-06be3d81b963","Type":"ContainerDied","Data":"c56db2f1c6f7d67c272a1235b8914e89669824b3796414f355ec5eb8bad20cfd"} Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.687095 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" event={"ID":"828b46a2-0bd7-430c-8cd1-06be3d81b963","Type":"ContainerStarted","Data":"bbab8cbc6f7325b55b38ae3989ecccd986b3855b9320754a71a4631d8a6dbba7"} Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.699711 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-rd5pw" event={"ID":"f84522a7-f356-40c9-8ea0-13b1503993fe","Type":"ContainerStarted","Data":"7d8c16c52256a7f940166ce1567aa8686f167a0d8de2b989b4302e9a7cb4adf9"} Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.714399 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.716398 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75bb4695fc-zpn6h" event={"ID":"daf1dc83-5f20-4f49-8327-8d0bd5e11bc0","Type":"ContainerDied","Data":"e044cf7043e751ff88f4d9241c3e7ef008a001acbb71993f9c933a77386fe1dd"} Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.716912 4818 scope.go:117] "RemoveContainer" containerID="745b8b5a96456ef61afb5180857d09a32abeac9d1af1974e16409d1abf69594e" Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.811296 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-rd5pw" podStartSLOduration=2.8112781780000002 podStartE2EDuration="2.811278178s" podCreationTimestamp="2025-11-22 05:10:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:10:45.743943218 +0000 UTC m=+1398.318359775" watchObservedRunningTime="2025-11-22 05:10:45.811278178 +0000 UTC m=+1398.385694705" Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.839601 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75bb4695fc-zpn6h"] Nov 22 05:10:45 crc kubenswrapper[4818]: I1122 05:10:45.845621 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75bb4695fc-zpn6h"] Nov 22 05:10:46 crc kubenswrapper[4818]: I1122 05:10:46.306590 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daf1dc83-5f20-4f49-8327-8d0bd5e11bc0" path="/var/lib/kubelet/pods/daf1dc83-5f20-4f49-8327-8d0bd5e11bc0/volumes" Nov 22 05:10:46 crc kubenswrapper[4818]: I1122 05:10:46.731014 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" event={"ID":"828b46a2-0bd7-430c-8cd1-06be3d81b963","Type":"ContainerStarted","Data":"a283aefa9ac79d3f17223b04c4310fb68a6c2e743bda42644cccace7cc2d369e"} Nov 22 05:10:46 crc kubenswrapper[4818]: I1122 05:10:46.731471 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:10:46 crc kubenswrapper[4818]: I1122 05:10:46.757785 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" podStartSLOduration=3.7577656619999997 podStartE2EDuration="3.757765662s" podCreationTimestamp="2025-11-22 05:10:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:10:46.749393209 +0000 UTC m=+1399.323809986" watchObservedRunningTime="2025-11-22 05:10:46.757765662 +0000 UTC m=+1399.332182199" Nov 22 05:10:47 crc kubenswrapper[4818]: I1122 05:10:47.752965 4818 generic.go:334] "Generic (PLEG): container finished" podID="486177c7-b0c3-45ba-a18c-79509154c9df" containerID="566e1df2e1e8ff694af911b80125bc51fa9ebcaae0be783202453e0a0427e77f" exitCode=0 Nov 22 05:10:47 crc kubenswrapper[4818]: I1122 05:10:47.753067 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jpkvr" event={"ID":"486177c7-b0c3-45ba-a18c-79509154c9df","Type":"ContainerDied","Data":"566e1df2e1e8ff694af911b80125bc51fa9ebcaae0be783202453e0a0427e77f"} Nov 22 05:10:51 crc kubenswrapper[4818]: I1122 05:10:51.265020 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:10:51 crc kubenswrapper[4818]: I1122 05:10:51.265705 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:10:52 crc kubenswrapper[4818]: I1122 05:10:52.792623 4818 generic.go:334] "Generic (PLEG): container finished" podID="54ceda18-1bdf-4adf-bb72-3c323f693eb0" containerID="f7a4d0ea11175f50c5e59c18c7fb82d32bcd64c49e2e4b60395c59ecb1aa066d" exitCode=0 Nov 22 05:10:52 crc kubenswrapper[4818]: I1122 05:10:52.792685 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t5p4f" event={"ID":"54ceda18-1bdf-4adf-bb72-3c323f693eb0","Type":"ContainerDied","Data":"f7a4d0ea11175f50c5e59c18c7fb82d32bcd64c49e2e4b60395c59ecb1aa066d"} Nov 22 05:10:53 crc kubenswrapper[4818]: I1122 05:10:53.337716 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jpkvr" Nov 22 05:10:53 crc kubenswrapper[4818]: I1122 05:10:53.501985 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/486177c7-b0c3-45ba-a18c-79509154c9df-db-sync-config-data\") pod \"486177c7-b0c3-45ba-a18c-79509154c9df\" (UID: \"486177c7-b0c3-45ba-a18c-79509154c9df\") " Nov 22 05:10:53 crc kubenswrapper[4818]: I1122 05:10:53.502145 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/486177c7-b0c3-45ba-a18c-79509154c9df-config-data\") pod \"486177c7-b0c3-45ba-a18c-79509154c9df\" (UID: \"486177c7-b0c3-45ba-a18c-79509154c9df\") " Nov 22 05:10:53 crc kubenswrapper[4818]: I1122 05:10:53.502198 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-br247\" (UniqueName: \"kubernetes.io/projected/486177c7-b0c3-45ba-a18c-79509154c9df-kube-api-access-br247\") pod \"486177c7-b0c3-45ba-a18c-79509154c9df\" (UID: \"486177c7-b0c3-45ba-a18c-79509154c9df\") " Nov 22 05:10:53 crc kubenswrapper[4818]: I1122 05:10:53.502361 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/486177c7-b0c3-45ba-a18c-79509154c9df-combined-ca-bundle\") pod \"486177c7-b0c3-45ba-a18c-79509154c9df\" (UID: \"486177c7-b0c3-45ba-a18c-79509154c9df\") " Nov 22 05:10:53 crc kubenswrapper[4818]: I1122 05:10:53.507917 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/486177c7-b0c3-45ba-a18c-79509154c9df-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "486177c7-b0c3-45ba-a18c-79509154c9df" (UID: "486177c7-b0c3-45ba-a18c-79509154c9df"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:10:53 crc kubenswrapper[4818]: I1122 05:10:53.507930 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/486177c7-b0c3-45ba-a18c-79509154c9df-kube-api-access-br247" (OuterVolumeSpecName: "kube-api-access-br247") pod "486177c7-b0c3-45ba-a18c-79509154c9df" (UID: "486177c7-b0c3-45ba-a18c-79509154c9df"). InnerVolumeSpecName "kube-api-access-br247". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:10:53 crc kubenswrapper[4818]: I1122 05:10:53.531368 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/486177c7-b0c3-45ba-a18c-79509154c9df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "486177c7-b0c3-45ba-a18c-79509154c9df" (UID: "486177c7-b0c3-45ba-a18c-79509154c9df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:10:53 crc kubenswrapper[4818]: I1122 05:10:53.548326 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/486177c7-b0c3-45ba-a18c-79509154c9df-config-data" (OuterVolumeSpecName: "config-data") pod "486177c7-b0c3-45ba-a18c-79509154c9df" (UID: "486177c7-b0c3-45ba-a18c-79509154c9df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:10:53 crc kubenswrapper[4818]: I1122 05:10:53.604126 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/486177c7-b0c3-45ba-a18c-79509154c9df-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:53 crc kubenswrapper[4818]: I1122 05:10:53.604169 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-br247\" (UniqueName: \"kubernetes.io/projected/486177c7-b0c3-45ba-a18c-79509154c9df-kube-api-access-br247\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:53 crc kubenswrapper[4818]: I1122 05:10:53.604184 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/486177c7-b0c3-45ba-a18c-79509154c9df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:53 crc kubenswrapper[4818]: I1122 05:10:53.604197 4818 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/486177c7-b0c3-45ba-a18c-79509154c9df-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:10:53 crc kubenswrapper[4818]: I1122 05:10:53.803163 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jpkvr" Nov 22 05:10:53 crc kubenswrapper[4818]: I1122 05:10:53.803158 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jpkvr" event={"ID":"486177c7-b0c3-45ba-a18c-79509154c9df","Type":"ContainerDied","Data":"41272fc91993fc39171c161d477bebda06b770971e966d3727dbcc6e9d6392a7"} Nov 22 05:10:53 crc kubenswrapper[4818]: I1122 05:10:53.803217 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41272fc91993fc39171c161d477bebda06b770971e966d3727dbcc6e9d6392a7" Nov 22 05:10:53 crc kubenswrapper[4818]: I1122 05:10:53.826045 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:10:53 crc kubenswrapper[4818]: I1122 05:10:53.899175 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kfb5r"] Nov 22 05:10:53 crc kubenswrapper[4818]: I1122 05:10:53.899447 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" podUID="044b1787-72fc-46a3-b73b-fb412ebc4a28" containerName="dnsmasq-dns" containerID="cri-o://92f506c62a4fd3211edc2e9d37d390268922b279faaec8732889f20b9cdcdd78" gracePeriod=10 Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.720333 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-jlc6c"] Nov 22 05:10:54 crc kubenswrapper[4818]: E1122 05:10:54.720875 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daf1dc83-5f20-4f49-8327-8d0bd5e11bc0" containerName="init" Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.720892 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="daf1dc83-5f20-4f49-8327-8d0bd5e11bc0" containerName="init" Nov 22 05:10:54 crc kubenswrapper[4818]: E1122 05:10:54.720911 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="486177c7-b0c3-45ba-a18c-79509154c9df" containerName="glance-db-sync" Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.720919 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="486177c7-b0c3-45ba-a18c-79509154c9df" containerName="glance-db-sync" Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.721072 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="486177c7-b0c3-45ba-a18c-79509154c9df" containerName="glance-db-sync" Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.721084 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="daf1dc83-5f20-4f49-8327-8d0bd5e11bc0" containerName="init" Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.721886 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.743067 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-jlc6c"] Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.819399 4818 generic.go:334] "Generic (PLEG): container finished" podID="044b1787-72fc-46a3-b73b-fb412ebc4a28" containerID="92f506c62a4fd3211edc2e9d37d390268922b279faaec8732889f20b9cdcdd78" exitCode=0 Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.819440 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" event={"ID":"044b1787-72fc-46a3-b73b-fb412ebc4a28","Type":"ContainerDied","Data":"92f506c62a4fd3211edc2e9d37d390268922b279faaec8732889f20b9cdcdd78"} Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.842565 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-ovsdbserver-nb\") pod \"dnsmasq-dns-7987f74bbc-jlc6c\" (UID: \"86b5909b-ca6a-4251-a03a-d0accd183c3a\") " pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.842630 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-ovsdbserver-sb\") pod \"dnsmasq-dns-7987f74bbc-jlc6c\" (UID: \"86b5909b-ca6a-4251-a03a-d0accd183c3a\") " pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.842651 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-dns-svc\") pod \"dnsmasq-dns-7987f74bbc-jlc6c\" (UID: \"86b5909b-ca6a-4251-a03a-d0accd183c3a\") " pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.842725 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-config\") pod \"dnsmasq-dns-7987f74bbc-jlc6c\" (UID: \"86b5909b-ca6a-4251-a03a-d0accd183c3a\") " pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.842748 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r99p2\" (UniqueName: \"kubernetes.io/projected/86b5909b-ca6a-4251-a03a-d0accd183c3a-kube-api-access-r99p2\") pod \"dnsmasq-dns-7987f74bbc-jlc6c\" (UID: \"86b5909b-ca6a-4251-a03a-d0accd183c3a\") " pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.943774 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-config\") pod \"dnsmasq-dns-7987f74bbc-jlc6c\" (UID: \"86b5909b-ca6a-4251-a03a-d0accd183c3a\") " pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.943818 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r99p2\" (UniqueName: \"kubernetes.io/projected/86b5909b-ca6a-4251-a03a-d0accd183c3a-kube-api-access-r99p2\") pod \"dnsmasq-dns-7987f74bbc-jlc6c\" (UID: \"86b5909b-ca6a-4251-a03a-d0accd183c3a\") " pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.943889 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-ovsdbserver-nb\") pod \"dnsmasq-dns-7987f74bbc-jlc6c\" (UID: \"86b5909b-ca6a-4251-a03a-d0accd183c3a\") " pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.943925 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-ovsdbserver-sb\") pod \"dnsmasq-dns-7987f74bbc-jlc6c\" (UID: \"86b5909b-ca6a-4251-a03a-d0accd183c3a\") " pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.943948 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-dns-svc\") pod \"dnsmasq-dns-7987f74bbc-jlc6c\" (UID: \"86b5909b-ca6a-4251-a03a-d0accd183c3a\") " pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.944984 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-dns-svc\") pod \"dnsmasq-dns-7987f74bbc-jlc6c\" (UID: \"86b5909b-ca6a-4251-a03a-d0accd183c3a\") " pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.945320 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-config\") pod \"dnsmasq-dns-7987f74bbc-jlc6c\" (UID: \"86b5909b-ca6a-4251-a03a-d0accd183c3a\") " pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.945528 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-ovsdbserver-nb\") pod \"dnsmasq-dns-7987f74bbc-jlc6c\" (UID: \"86b5909b-ca6a-4251-a03a-d0accd183c3a\") " pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.945797 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-ovsdbserver-sb\") pod \"dnsmasq-dns-7987f74bbc-jlc6c\" (UID: \"86b5909b-ca6a-4251-a03a-d0accd183c3a\") " pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:10:54 crc kubenswrapper[4818]: I1122 05:10:54.986518 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r99p2\" (UniqueName: \"kubernetes.io/projected/86b5909b-ca6a-4251-a03a-d0accd183c3a-kube-api-access-r99p2\") pod \"dnsmasq-dns-7987f74bbc-jlc6c\" (UID: \"86b5909b-ca6a-4251-a03a-d0accd183c3a\") " pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:10:55 crc kubenswrapper[4818]: I1122 05:10:55.074673 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:10:55 crc kubenswrapper[4818]: I1122 05:10:55.466963 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" podUID="044b1787-72fc-46a3-b73b-fb412ebc4a28" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.111:5353: connect: connection refused" Nov 22 05:10:57 crc kubenswrapper[4818]: E1122 05:10:57.809416 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Nov 22 05:10:57 crc kubenswrapper[4818]: E1122 05:10:57.811140 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n688h85h654h5fh89h6ch576h556h5b4hc4h5bhf6h544hd7h66dh8bh5ddhcdh558h585h66h668h648h76h595h77h599h55dh64ch66hd8h95q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x826k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(0f6203d4-9fff-483c-bc5c-84deb159d47d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 05:11:00 crc kubenswrapper[4818]: I1122 05:11:00.466843 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" podUID="044b1787-72fc-46a3-b73b-fb412ebc4a28" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.111:5353: connect: connection refused" Nov 22 05:11:05 crc kubenswrapper[4818]: I1122 05:11:05.467158 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" podUID="044b1787-72fc-46a3-b73b-fb412ebc4a28" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.111:5353: connect: connection refused" Nov 22 05:11:05 crc kubenswrapper[4818]: I1122 05:11:05.468131 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.513274 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.613667 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-credential-keys\") pod \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.614119 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-scripts\") pod \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.614154 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-combined-ca-bundle\") pod \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.614184 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-fernet-keys\") pod \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.614204 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92hgm\" (UniqueName: \"kubernetes.io/projected/54ceda18-1bdf-4adf-bb72-3c323f693eb0-kube-api-access-92hgm\") pod \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.614229 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-config-data\") pod \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\" (UID: \"54ceda18-1bdf-4adf-bb72-3c323f693eb0\") " Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.619524 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "54ceda18-1bdf-4adf-bb72-3c323f693eb0" (UID: "54ceda18-1bdf-4adf-bb72-3c323f693eb0"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.634954 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-scripts" (OuterVolumeSpecName: "scripts") pod "54ceda18-1bdf-4adf-bb72-3c323f693eb0" (UID: "54ceda18-1bdf-4adf-bb72-3c323f693eb0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.636362 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54ceda18-1bdf-4adf-bb72-3c323f693eb0-kube-api-access-92hgm" (OuterVolumeSpecName: "kube-api-access-92hgm") pod "54ceda18-1bdf-4adf-bb72-3c323f693eb0" (UID: "54ceda18-1bdf-4adf-bb72-3c323f693eb0"). InnerVolumeSpecName "kube-api-access-92hgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.636571 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "54ceda18-1bdf-4adf-bb72-3c323f693eb0" (UID: "54ceda18-1bdf-4adf-bb72-3c323f693eb0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.642713 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "54ceda18-1bdf-4adf-bb72-3c323f693eb0" (UID: "54ceda18-1bdf-4adf-bb72-3c323f693eb0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.650445 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-config-data" (OuterVolumeSpecName: "config-data") pod "54ceda18-1bdf-4adf-bb72-3c323f693eb0" (UID: "54ceda18-1bdf-4adf-bb72-3c323f693eb0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.716639 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.716701 4818 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.716716 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92hgm\" (UniqueName: \"kubernetes.io/projected/54ceda18-1bdf-4adf-bb72-3c323f693eb0-kube-api-access-92hgm\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.716731 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.716802 4818 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.716815 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54ceda18-1bdf-4adf-bb72-3c323f693eb0-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.938817 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t5p4f" event={"ID":"54ceda18-1bdf-4adf-bb72-3c323f693eb0","Type":"ContainerDied","Data":"88878bee1c87320de1d8d989b85291174dbbeecc3c30dafb3206e1c2dd0d45b5"} Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.938862 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88878bee1c87320de1d8d989b85291174dbbeecc3c30dafb3206e1c2dd0d45b5" Nov 22 05:11:08 crc kubenswrapper[4818]: I1122 05:11:08.938869 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t5p4f" Nov 22 05:11:09 crc kubenswrapper[4818]: E1122 05:11:09.024542 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Nov 22 05:11:09 crc kubenswrapper[4818]: E1122 05:11:09.024679 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-577sx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-n2ldz_openstack(2a8a8454-074a-48c1-b1f2-f800dd6f0f0a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 05:11:09 crc kubenswrapper[4818]: E1122 05:11:09.026025 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-n2ldz" podUID="2a8a8454-074a-48c1-b1f2-f800dd6f0f0a" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.590444 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-t5p4f"] Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.597005 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-t5p4f"] Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.687057 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-8mhxq"] Nov 22 05:11:09 crc kubenswrapper[4818]: E1122 05:11:09.688494 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54ceda18-1bdf-4adf-bb72-3c323f693eb0" containerName="keystone-bootstrap" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.688551 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="54ceda18-1bdf-4adf-bb72-3c323f693eb0" containerName="keystone-bootstrap" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.688761 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="54ceda18-1bdf-4adf-bb72-3c323f693eb0" containerName="keystone-bootstrap" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.689575 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.691923 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.691984 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.691920 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-tds7x" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.692277 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.692459 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.694165 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-8mhxq"] Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.728927 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-scripts\") pod \"keystone-bootstrap-8mhxq\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.728972 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-config-data\") pod \"keystone-bootstrap-8mhxq\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.729022 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-combined-ca-bundle\") pod \"keystone-bootstrap-8mhxq\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.729110 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-fernet-keys\") pod \"keystone-bootstrap-8mhxq\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.729127 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-credential-keys\") pod \"keystone-bootstrap-8mhxq\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.729279 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vl64\" (UniqueName: \"kubernetes.io/projected/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-kube-api-access-4vl64\") pod \"keystone-bootstrap-8mhxq\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.830512 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-fernet-keys\") pod \"keystone-bootstrap-8mhxq\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.830548 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-credential-keys\") pod \"keystone-bootstrap-8mhxq\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.830600 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vl64\" (UniqueName: \"kubernetes.io/projected/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-kube-api-access-4vl64\") pod \"keystone-bootstrap-8mhxq\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.830632 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-scripts\") pod \"keystone-bootstrap-8mhxq\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.830654 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-config-data\") pod \"keystone-bootstrap-8mhxq\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.830704 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-combined-ca-bundle\") pod \"keystone-bootstrap-8mhxq\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.840579 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-credential-keys\") pod \"keystone-bootstrap-8mhxq\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.840671 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-combined-ca-bundle\") pod \"keystone-bootstrap-8mhxq\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.840671 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-fernet-keys\") pod \"keystone-bootstrap-8mhxq\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.842051 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-config-data\") pod \"keystone-bootstrap-8mhxq\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.843618 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-scripts\") pod \"keystone-bootstrap-8mhxq\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:09 crc kubenswrapper[4818]: I1122 05:11:09.853334 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vl64\" (UniqueName: \"kubernetes.io/projected/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-kube-api-access-4vl64\") pod \"keystone-bootstrap-8mhxq\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:09 crc kubenswrapper[4818]: E1122 05:11:09.984663 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-n2ldz" podUID="2a8a8454-074a-48c1-b1f2-f800dd6f0f0a" Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.012506 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:10 crc kubenswrapper[4818]: E1122 05:11:10.170839 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 22 05:11:10 crc kubenswrapper[4818]: E1122 05:11:10.170981 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nwvrh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-7spb7_openstack(0ca1db80-7d8d-44cf-80bf-afa1c0798182): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 05:11:10 crc kubenswrapper[4818]: E1122 05:11:10.172373 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-7spb7" podUID="0ca1db80-7d8d-44cf-80bf-afa1c0798182" Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.233852 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.308952 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54ceda18-1bdf-4adf-bb72-3c323f693eb0" path="/var/lib/kubelet/pods/54ceda18-1bdf-4adf-bb72-3c323f693eb0/volumes" Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.338670 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-config\") pod \"044b1787-72fc-46a3-b73b-fb412ebc4a28\" (UID: \"044b1787-72fc-46a3-b73b-fb412ebc4a28\") " Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.338822 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-ovsdbserver-nb\") pod \"044b1787-72fc-46a3-b73b-fb412ebc4a28\" (UID: \"044b1787-72fc-46a3-b73b-fb412ebc4a28\") " Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.338894 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-dns-svc\") pod \"044b1787-72fc-46a3-b73b-fb412ebc4a28\" (UID: \"044b1787-72fc-46a3-b73b-fb412ebc4a28\") " Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.338923 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwfn8\" (UniqueName: \"kubernetes.io/projected/044b1787-72fc-46a3-b73b-fb412ebc4a28-kube-api-access-bwfn8\") pod \"044b1787-72fc-46a3-b73b-fb412ebc4a28\" (UID: \"044b1787-72fc-46a3-b73b-fb412ebc4a28\") " Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.339012 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-ovsdbserver-sb\") pod \"044b1787-72fc-46a3-b73b-fb412ebc4a28\" (UID: \"044b1787-72fc-46a3-b73b-fb412ebc4a28\") " Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.347509 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/044b1787-72fc-46a3-b73b-fb412ebc4a28-kube-api-access-bwfn8" (OuterVolumeSpecName: "kube-api-access-bwfn8") pod "044b1787-72fc-46a3-b73b-fb412ebc4a28" (UID: "044b1787-72fc-46a3-b73b-fb412ebc4a28"). InnerVolumeSpecName "kube-api-access-bwfn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.382376 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "044b1787-72fc-46a3-b73b-fb412ebc4a28" (UID: "044b1787-72fc-46a3-b73b-fb412ebc4a28"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.385232 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-config" (OuterVolumeSpecName: "config") pod "044b1787-72fc-46a3-b73b-fb412ebc4a28" (UID: "044b1787-72fc-46a3-b73b-fb412ebc4a28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.389060 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "044b1787-72fc-46a3-b73b-fb412ebc4a28" (UID: "044b1787-72fc-46a3-b73b-fb412ebc4a28"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.394956 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "044b1787-72fc-46a3-b73b-fb412ebc4a28" (UID: "044b1787-72fc-46a3-b73b-fb412ebc4a28"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.441008 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.441042 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.441055 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.441063 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/044b1787-72fc-46a3-b73b-fb412ebc4a28-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.441072 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwfn8\" (UniqueName: \"kubernetes.io/projected/044b1787-72fc-46a3-b73b-fb412ebc4a28-kube-api-access-bwfn8\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.766321 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-jlc6c"] Nov 22 05:11:10 crc kubenswrapper[4818]: W1122 05:11:10.776891 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86b5909b_ca6a_4251_a03a_d0accd183c3a.slice/crio-5af9564cc8115832f57190aa122059843a317340f9f7480c200ba2db28712956 WatchSource:0}: Error finding container 5af9564cc8115832f57190aa122059843a317340f9f7480c200ba2db28712956: Status 404 returned error can't find the container with id 5af9564cc8115832f57190aa122059843a317340f9f7480c200ba2db28712956 Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.895816 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-8mhxq"] Nov 22 05:11:10 crc kubenswrapper[4818]: W1122 05:11:10.897218 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a3127b6_3ea9_4c8a_aa23_e1e0452d3aad.slice/crio-719e1133811f230f8a13fd8d3c042f4c799f231cc0e00bb610b3236e17acbd6e WatchSource:0}: Error finding container 719e1133811f230f8a13fd8d3c042f4c799f231cc0e00bb610b3236e17acbd6e: Status 404 returned error can't find the container with id 719e1133811f230f8a13fd8d3c042f4c799f231cc0e00bb610b3236e17acbd6e Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.954986 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" event={"ID":"044b1787-72fc-46a3-b73b-fb412ebc4a28","Type":"ContainerDied","Data":"f4589124ace1bb63071aa23a6df309ac67df4dda79dbd8776f65796a84f947d4"} Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.955044 4818 scope.go:117] "RemoveContainer" containerID="92f506c62a4fd3211edc2e9d37d390268922b279faaec8732889f20b9cdcdd78" Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.955001 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-kfb5r" Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.958160 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8mhxq" event={"ID":"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad","Type":"ContainerStarted","Data":"719e1133811f230f8a13fd8d3c042f4c799f231cc0e00bb610b3236e17acbd6e"} Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.960032 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qnwqb" event={"ID":"e86e2277-a801-4562-8741-84420e0d8b5e","Type":"ContainerStarted","Data":"52103b4c483d95017e34149bc547bcb135efd0fe9d874cd4ef26bdfe8dedd602"} Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.967112 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" event={"ID":"86b5909b-ca6a-4251-a03a-d0accd183c3a","Type":"ContainerStarted","Data":"5af9564cc8115832f57190aa122059843a317340f9f7480c200ba2db28712956"} Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.969196 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f6203d4-9fff-483c-bc5c-84deb159d47d","Type":"ContainerStarted","Data":"3c7694200619b732d5d71a5c66213d99c4dc98ba443c3669d13766ec94e389b6"} Nov 22 05:11:10 crc kubenswrapper[4818]: E1122 05:11:10.969892 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-7spb7" podUID="0ca1db80-7d8d-44cf-80bf-afa1c0798182" Nov 22 05:11:10 crc kubenswrapper[4818]: I1122 05:11:10.989097 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-qnwqb" podStartSLOduration=2.424860954 podStartE2EDuration="27.989072546s" podCreationTimestamp="2025-11-22 05:10:43 +0000 UTC" firstStartedPulling="2025-11-22 05:10:44.582698127 +0000 UTC m=+1397.157114654" lastFinishedPulling="2025-11-22 05:11:10.146909719 +0000 UTC m=+1422.721326246" observedRunningTime="2025-11-22 05:11:10.978975139 +0000 UTC m=+1423.553391666" watchObservedRunningTime="2025-11-22 05:11:10.989072546 +0000 UTC m=+1423.563489073" Nov 22 05:11:11 crc kubenswrapper[4818]: I1122 05:11:11.008664 4818 scope.go:117] "RemoveContainer" containerID="de1bf831a88a2bc5fd75be68335d24ca6ec6611149ccb8c5f88ca9b55dd3e3f4" Nov 22 05:11:11 crc kubenswrapper[4818]: I1122 05:11:11.068347 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kfb5r"] Nov 22 05:11:11 crc kubenswrapper[4818]: I1122 05:11:11.076395 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kfb5r"] Nov 22 05:11:11 crc kubenswrapper[4818]: I1122 05:11:11.983158 4818 generic.go:334] "Generic (PLEG): container finished" podID="86b5909b-ca6a-4251-a03a-d0accd183c3a" containerID="00b89d651bd7219dd8a3927dc06d45b661001ad95cba245901d5d110d9250041" exitCode=0 Nov 22 05:11:11 crc kubenswrapper[4818]: I1122 05:11:11.983278 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" event={"ID":"86b5909b-ca6a-4251-a03a-d0accd183c3a","Type":"ContainerDied","Data":"00b89d651bd7219dd8a3927dc06d45b661001ad95cba245901d5d110d9250041"} Nov 22 05:11:11 crc kubenswrapper[4818]: I1122 05:11:11.991118 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8mhxq" event={"ID":"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad","Type":"ContainerStarted","Data":"a4cb577a918c0af02b86283e9fe25886815e441e60bfdf13c49010b787d99a52"} Nov 22 05:11:12 crc kubenswrapper[4818]: I1122 05:11:12.040164 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-8mhxq" podStartSLOduration=3.04014147 podStartE2EDuration="3.04014147s" podCreationTimestamp="2025-11-22 05:11:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:11:12.032198951 +0000 UTC m=+1424.606615508" watchObservedRunningTime="2025-11-22 05:11:12.04014147 +0000 UTC m=+1424.614557997" Nov 22 05:11:12 crc kubenswrapper[4818]: I1122 05:11:12.301235 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="044b1787-72fc-46a3-b73b-fb412ebc4a28" path="/var/lib/kubelet/pods/044b1787-72fc-46a3-b73b-fb412ebc4a28/volumes" Nov 22 05:11:13 crc kubenswrapper[4818]: I1122 05:11:13.000189 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" event={"ID":"86b5909b-ca6a-4251-a03a-d0accd183c3a","Type":"ContainerStarted","Data":"8b389f86c5110e71cbf7f20b4806d0a801afb4bd669d562a6283584396f1dd38"} Nov 22 05:11:13 crc kubenswrapper[4818]: I1122 05:11:13.023605 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" podStartSLOduration=19.023588132 podStartE2EDuration="19.023588132s" podCreationTimestamp="2025-11-22 05:10:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:11:13.021270051 +0000 UTC m=+1425.595686568" watchObservedRunningTime="2025-11-22 05:11:13.023588132 +0000 UTC m=+1425.598004659" Nov 22 05:11:14 crc kubenswrapper[4818]: I1122 05:11:14.008125 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:11:15 crc kubenswrapper[4818]: I1122 05:11:15.015320 4818 generic.go:334] "Generic (PLEG): container finished" podID="8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad" containerID="a4cb577a918c0af02b86283e9fe25886815e441e60bfdf13c49010b787d99a52" exitCode=0 Nov 22 05:11:15 crc kubenswrapper[4818]: I1122 05:11:15.015375 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8mhxq" event={"ID":"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad","Type":"ContainerDied","Data":"a4cb577a918c0af02b86283e9fe25886815e441e60bfdf13c49010b787d99a52"} Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.026547 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f6203d4-9fff-483c-bc5c-84deb159d47d","Type":"ContainerStarted","Data":"7cb97b13bbe6cb54623a632b4f7c9b161afb6c86f35e9e141fd7166dc28a44e2"} Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.028306 4818 generic.go:334] "Generic (PLEG): container finished" podID="e86e2277-a801-4562-8741-84420e0d8b5e" containerID="52103b4c483d95017e34149bc547bcb135efd0fe9d874cd4ef26bdfe8dedd602" exitCode=0 Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.028620 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qnwqb" event={"ID":"e86e2277-a801-4562-8741-84420e0d8b5e","Type":"ContainerDied","Data":"52103b4c483d95017e34149bc547bcb135efd0fe9d874cd4ef26bdfe8dedd602"} Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.377413 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.559381 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-scripts\") pod \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.559676 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-config-data\") pod \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.559721 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-combined-ca-bundle\") pod \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.559822 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-credential-keys\") pod \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.559907 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vl64\" (UniqueName: \"kubernetes.io/projected/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-kube-api-access-4vl64\") pod \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.559975 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-fernet-keys\") pod \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\" (UID: \"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad\") " Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.566814 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad" (UID: "8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.567221 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad" (UID: "8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.567897 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-scripts" (OuterVolumeSpecName: "scripts") pod "8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad" (UID: "8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.572990 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-kube-api-access-4vl64" (OuterVolumeSpecName: "kube-api-access-4vl64") pod "8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad" (UID: "8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad"). InnerVolumeSpecName "kube-api-access-4vl64". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.591996 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad" (UID: "8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.592052 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-config-data" (OuterVolumeSpecName: "config-data") pod "8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad" (UID: "8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.661624 4818 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.661658 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vl64\" (UniqueName: \"kubernetes.io/projected/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-kube-api-access-4vl64\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.661672 4818 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.661688 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.661699 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:16 crc kubenswrapper[4818]: I1122 05:11:16.661709 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.041378 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8mhxq" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.041820 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8mhxq" event={"ID":"8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad","Type":"ContainerDied","Data":"719e1133811f230f8a13fd8d3c042f4c799f231cc0e00bb610b3236e17acbd6e"} Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.041872 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="719e1133811f230f8a13fd8d3c042f4c799f231cc0e00bb610b3236e17acbd6e" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.157517 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-54d7c687fd-74zvh"] Nov 22 05:11:17 crc kubenswrapper[4818]: E1122 05:11:17.158007 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="044b1787-72fc-46a3-b73b-fb412ebc4a28" containerName="dnsmasq-dns" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.158031 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="044b1787-72fc-46a3-b73b-fb412ebc4a28" containerName="dnsmasq-dns" Nov 22 05:11:17 crc kubenswrapper[4818]: E1122 05:11:17.158046 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad" containerName="keystone-bootstrap" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.158056 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad" containerName="keystone-bootstrap" Nov 22 05:11:17 crc kubenswrapper[4818]: E1122 05:11:17.158082 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="044b1787-72fc-46a3-b73b-fb412ebc4a28" containerName="init" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.158103 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="044b1787-72fc-46a3-b73b-fb412ebc4a28" containerName="init" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.158327 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad" containerName="keystone-bootstrap" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.158353 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="044b1787-72fc-46a3-b73b-fb412ebc4a28" containerName="dnsmasq-dns" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.159004 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.162696 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-tds7x" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.162905 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.163062 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.163219 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.163417 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.172853 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.184345 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-54d7c687fd-74zvh"] Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.273388 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-config-data\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.273451 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-scripts\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.273493 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-fernet-keys\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.273542 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-public-tls-certs\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.273571 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-internal-tls-certs\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.273599 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-credential-keys\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.273635 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-combined-ca-bundle\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.273660 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vhqs\" (UniqueName: \"kubernetes.io/projected/fe562192-8025-4265-826d-6e7b1400b281-kube-api-access-4vhqs\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.380702 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-config-data\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.381177 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-scripts\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.381268 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-fernet-keys\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.381393 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-public-tls-certs\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.381451 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-internal-tls-certs\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.381504 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-credential-keys\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.381611 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-combined-ca-bundle\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.381659 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vhqs\" (UniqueName: \"kubernetes.io/projected/fe562192-8025-4265-826d-6e7b1400b281-kube-api-access-4vhqs\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.384519 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-scripts\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.385783 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-config-data\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.386023 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-internal-tls-certs\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.386199 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-fernet-keys\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.387670 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-combined-ca-bundle\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.390808 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-credential-keys\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.397512 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe562192-8025-4265-826d-6e7b1400b281-public-tls-certs\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.405906 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vhqs\" (UniqueName: \"kubernetes.io/projected/fe562192-8025-4265-826d-6e7b1400b281-kube-api-access-4vhqs\") pod \"keystone-54d7c687fd-74zvh\" (UID: \"fe562192-8025-4265-826d-6e7b1400b281\") " pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.475745 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.477204 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qnwqb" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.590004 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzf5g\" (UniqueName: \"kubernetes.io/projected/e86e2277-a801-4562-8741-84420e0d8b5e-kube-api-access-gzf5g\") pod \"e86e2277-a801-4562-8741-84420e0d8b5e\" (UID: \"e86e2277-a801-4562-8741-84420e0d8b5e\") " Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.590068 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e86e2277-a801-4562-8741-84420e0d8b5e-logs\") pod \"e86e2277-a801-4562-8741-84420e0d8b5e\" (UID: \"e86e2277-a801-4562-8741-84420e0d8b5e\") " Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.590157 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86e2277-a801-4562-8741-84420e0d8b5e-combined-ca-bundle\") pod \"e86e2277-a801-4562-8741-84420e0d8b5e\" (UID: \"e86e2277-a801-4562-8741-84420e0d8b5e\") " Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.590216 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e86e2277-a801-4562-8741-84420e0d8b5e-config-data\") pod \"e86e2277-a801-4562-8741-84420e0d8b5e\" (UID: \"e86e2277-a801-4562-8741-84420e0d8b5e\") " Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.590283 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e86e2277-a801-4562-8741-84420e0d8b5e-scripts\") pod \"e86e2277-a801-4562-8741-84420e0d8b5e\" (UID: \"e86e2277-a801-4562-8741-84420e0d8b5e\") " Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.591900 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e86e2277-a801-4562-8741-84420e0d8b5e-logs" (OuterVolumeSpecName: "logs") pod "e86e2277-a801-4562-8741-84420e0d8b5e" (UID: "e86e2277-a801-4562-8741-84420e0d8b5e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.594852 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e86e2277-a801-4562-8741-84420e0d8b5e-kube-api-access-gzf5g" (OuterVolumeSpecName: "kube-api-access-gzf5g") pod "e86e2277-a801-4562-8741-84420e0d8b5e" (UID: "e86e2277-a801-4562-8741-84420e0d8b5e"). InnerVolumeSpecName "kube-api-access-gzf5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.605436 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e86e2277-a801-4562-8741-84420e0d8b5e-scripts" (OuterVolumeSpecName: "scripts") pod "e86e2277-a801-4562-8741-84420e0d8b5e" (UID: "e86e2277-a801-4562-8741-84420e0d8b5e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.666469 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e86e2277-a801-4562-8741-84420e0d8b5e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e86e2277-a801-4562-8741-84420e0d8b5e" (UID: "e86e2277-a801-4562-8741-84420e0d8b5e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.676987 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e86e2277-a801-4562-8741-84420e0d8b5e-config-data" (OuterVolumeSpecName: "config-data") pod "e86e2277-a801-4562-8741-84420e0d8b5e" (UID: "e86e2277-a801-4562-8741-84420e0d8b5e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.692297 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzf5g\" (UniqueName: \"kubernetes.io/projected/e86e2277-a801-4562-8741-84420e0d8b5e-kube-api-access-gzf5g\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.692662 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e86e2277-a801-4562-8741-84420e0d8b5e-logs\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.692681 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86e2277-a801-4562-8741-84420e0d8b5e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.692692 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e86e2277-a801-4562-8741-84420e0d8b5e-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:17 crc kubenswrapper[4818]: I1122 05:11:17.692704 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e86e2277-a801-4562-8741-84420e0d8b5e-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.049572 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-54d7c687fd-74zvh"] Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.051976 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qnwqb" event={"ID":"e86e2277-a801-4562-8741-84420e0d8b5e","Type":"ContainerDied","Data":"f0a234acc6ddcc6853ec1698a5a3e2b505688f4b92adc73f19d17e7e5605ca66"} Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.052010 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0a234acc6ddcc6853ec1698a5a3e2b505688f4b92adc73f19d17e7e5605ca66" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.052062 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qnwqb" Nov 22 05:11:18 crc kubenswrapper[4818]: W1122 05:11:18.052188 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe562192_8025_4265_826d_6e7b1400b281.slice/crio-62281a53acec3999c11c5b5cfe8b4b95ae62532a65666490d0af9df43089603a WatchSource:0}: Error finding container 62281a53acec3999c11c5b5cfe8b4b95ae62532a65666490d0af9df43089603a: Status 404 returned error can't find the container with id 62281a53acec3999c11c5b5cfe8b4b95ae62532a65666490d0af9df43089603a Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.224198 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7b464767cb-jcwc8"] Nov 22 05:11:18 crc kubenswrapper[4818]: E1122 05:11:18.224561 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e86e2277-a801-4562-8741-84420e0d8b5e" containerName="placement-db-sync" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.224580 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="e86e2277-a801-4562-8741-84420e0d8b5e" containerName="placement-db-sync" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.224735 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="e86e2277-a801-4562-8741-84420e0d8b5e" containerName="placement-db-sync" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.236451 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7b464767cb-jcwc8"] Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.236548 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.238339 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.238983 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.239201 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-nnwmw" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.239350 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.239538 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.405030 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e48ac54-5e91-4747-b9e9-28aeb21180fd-combined-ca-bundle\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.405298 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rbth\" (UniqueName: \"kubernetes.io/projected/4e48ac54-5e91-4747-b9e9-28aeb21180fd-kube-api-access-7rbth\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.405323 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e48ac54-5e91-4747-b9e9-28aeb21180fd-config-data\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.405339 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e48ac54-5e91-4747-b9e9-28aeb21180fd-scripts\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.405357 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e48ac54-5e91-4747-b9e9-28aeb21180fd-internal-tls-certs\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.405411 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e48ac54-5e91-4747-b9e9-28aeb21180fd-public-tls-certs\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.405450 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e48ac54-5e91-4747-b9e9-28aeb21180fd-logs\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.507856 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e48ac54-5e91-4747-b9e9-28aeb21180fd-combined-ca-bundle\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.507910 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rbth\" (UniqueName: \"kubernetes.io/projected/4e48ac54-5e91-4747-b9e9-28aeb21180fd-kube-api-access-7rbth\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.507945 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e48ac54-5e91-4747-b9e9-28aeb21180fd-config-data\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.507969 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e48ac54-5e91-4747-b9e9-28aeb21180fd-scripts\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.507991 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e48ac54-5e91-4747-b9e9-28aeb21180fd-internal-tls-certs\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.508039 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e48ac54-5e91-4747-b9e9-28aeb21180fd-public-tls-certs\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.508080 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e48ac54-5e91-4747-b9e9-28aeb21180fd-logs\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.508655 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e48ac54-5e91-4747-b9e9-28aeb21180fd-logs\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.511681 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e48ac54-5e91-4747-b9e9-28aeb21180fd-config-data\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.511987 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e48ac54-5e91-4747-b9e9-28aeb21180fd-combined-ca-bundle\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.512104 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e48ac54-5e91-4747-b9e9-28aeb21180fd-scripts\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.512359 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e48ac54-5e91-4747-b9e9-28aeb21180fd-internal-tls-certs\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.514846 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e48ac54-5e91-4747-b9e9-28aeb21180fd-public-tls-certs\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.530884 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rbth\" (UniqueName: \"kubernetes.io/projected/4e48ac54-5e91-4747-b9e9-28aeb21180fd-kube-api-access-7rbth\") pod \"placement-7b464767cb-jcwc8\" (UID: \"4e48ac54-5e91-4747-b9e9-28aeb21180fd\") " pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.559570 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:18 crc kubenswrapper[4818]: I1122 05:11:18.998422 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7b464767cb-jcwc8"] Nov 22 05:11:19 crc kubenswrapper[4818]: W1122 05:11:19.001517 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e48ac54_5e91_4747_b9e9_28aeb21180fd.slice/crio-07ee95cc8a9855c05dfb1b0ac71411a0ba2343e2626724c20ba1efcb25b03299 WatchSource:0}: Error finding container 07ee95cc8a9855c05dfb1b0ac71411a0ba2343e2626724c20ba1efcb25b03299: Status 404 returned error can't find the container with id 07ee95cc8a9855c05dfb1b0ac71411a0ba2343e2626724c20ba1efcb25b03299 Nov 22 05:11:19 crc kubenswrapper[4818]: I1122 05:11:19.071432 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-54d7c687fd-74zvh" event={"ID":"fe562192-8025-4265-826d-6e7b1400b281","Type":"ContainerStarted","Data":"e64b70884d1e0ebb4a71233bdc750d18c66a3f6ade9f5661d115a54873dcc808"} Nov 22 05:11:19 crc kubenswrapper[4818]: I1122 05:11:19.071504 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-54d7c687fd-74zvh" event={"ID":"fe562192-8025-4265-826d-6e7b1400b281","Type":"ContainerStarted","Data":"62281a53acec3999c11c5b5cfe8b4b95ae62532a65666490d0af9df43089603a"} Nov 22 05:11:19 crc kubenswrapper[4818]: I1122 05:11:19.073093 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7b464767cb-jcwc8" event={"ID":"4e48ac54-5e91-4747-b9e9-28aeb21180fd","Type":"ContainerStarted","Data":"07ee95cc8a9855c05dfb1b0ac71411a0ba2343e2626724c20ba1efcb25b03299"} Nov 22 05:11:19 crc kubenswrapper[4818]: I1122 05:11:19.104480 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-54d7c687fd-74zvh" podStartSLOduration=2.104457092 podStartE2EDuration="2.104457092s" podCreationTimestamp="2025-11-22 05:11:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:11:19.096844181 +0000 UTC m=+1431.671260728" watchObservedRunningTime="2025-11-22 05:11:19.104457092 +0000 UTC m=+1431.678873629" Nov 22 05:11:20 crc kubenswrapper[4818]: I1122 05:11:20.076290 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:11:20 crc kubenswrapper[4818]: I1122 05:11:20.079238 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:20 crc kubenswrapper[4818]: I1122 05:11:20.124064 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-745b9ddc8c-cpmmw"] Nov 22 05:11:20 crc kubenswrapper[4818]: I1122 05:11:20.124618 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" podUID="828b46a2-0bd7-430c-8cd1-06be3d81b963" containerName="dnsmasq-dns" containerID="cri-o://a283aefa9ac79d3f17223b04c4310fb68a6c2e743bda42644cccace7cc2d369e" gracePeriod=10 Nov 22 05:11:21 crc kubenswrapper[4818]: I1122 05:11:21.094217 4818 generic.go:334] "Generic (PLEG): container finished" podID="828b46a2-0bd7-430c-8cd1-06be3d81b963" containerID="a283aefa9ac79d3f17223b04c4310fb68a6c2e743bda42644cccace7cc2d369e" exitCode=0 Nov 22 05:11:21 crc kubenswrapper[4818]: I1122 05:11:21.094312 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" event={"ID":"828b46a2-0bd7-430c-8cd1-06be3d81b963","Type":"ContainerDied","Data":"a283aefa9ac79d3f17223b04c4310fb68a6c2e743bda42644cccace7cc2d369e"} Nov 22 05:11:21 crc kubenswrapper[4818]: I1122 05:11:21.098498 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7b464767cb-jcwc8" event={"ID":"4e48ac54-5e91-4747-b9e9-28aeb21180fd","Type":"ContainerStarted","Data":"d5aa810711e5a1eda6e94d1af874b335d74cfc97021cb2f5a53b85adfcfbe326"} Nov 22 05:11:21 crc kubenswrapper[4818]: I1122 05:11:21.264774 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:11:21 crc kubenswrapper[4818]: I1122 05:11:21.264827 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:11:21 crc kubenswrapper[4818]: I1122 05:11:21.264864 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 05:11:21 crc kubenswrapper[4818]: I1122 05:11:21.265391 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f815386123126fabb0fbbe54df8d571044020c0a205a877246cd025618c83140"} pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 05:11:21 crc kubenswrapper[4818]: I1122 05:11:21.265435 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" containerID="cri-o://f815386123126fabb0fbbe54df8d571044020c0a205a877246cd025618c83140" gracePeriod=600 Nov 22 05:11:22 crc kubenswrapper[4818]: I1122 05:11:22.117742 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerDied","Data":"f815386123126fabb0fbbe54df8d571044020c0a205a877246cd025618c83140"} Nov 22 05:11:22 crc kubenswrapper[4818]: I1122 05:11:22.118052 4818 scope.go:117] "RemoveContainer" containerID="7f4e621532f88c0da5dac82557bd8694f552742e853952ce47999bc11fbbc049" Nov 22 05:11:22 crc kubenswrapper[4818]: I1122 05:11:22.117697 4818 generic.go:334] "Generic (PLEG): container finished" podID="0971c9ec-29ce-44eb-9875-ad4664143305" containerID="f815386123126fabb0fbbe54df8d571044020c0a205a877246cd025618c83140" exitCode=0 Nov 22 05:11:23 crc kubenswrapper[4818]: I1122 05:11:23.662678 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:11:23 crc kubenswrapper[4818]: I1122 05:11:23.803350 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-dns-svc\") pod \"828b46a2-0bd7-430c-8cd1-06be3d81b963\" (UID: \"828b46a2-0bd7-430c-8cd1-06be3d81b963\") " Nov 22 05:11:23 crc kubenswrapper[4818]: I1122 05:11:23.803465 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-config\") pod \"828b46a2-0bd7-430c-8cd1-06be3d81b963\" (UID: \"828b46a2-0bd7-430c-8cd1-06be3d81b963\") " Nov 22 05:11:23 crc kubenswrapper[4818]: I1122 05:11:23.803505 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-ovsdbserver-nb\") pod \"828b46a2-0bd7-430c-8cd1-06be3d81b963\" (UID: \"828b46a2-0bd7-430c-8cd1-06be3d81b963\") " Nov 22 05:11:23 crc kubenswrapper[4818]: I1122 05:11:23.803552 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxcvc\" (UniqueName: \"kubernetes.io/projected/828b46a2-0bd7-430c-8cd1-06be3d81b963-kube-api-access-rxcvc\") pod \"828b46a2-0bd7-430c-8cd1-06be3d81b963\" (UID: \"828b46a2-0bd7-430c-8cd1-06be3d81b963\") " Nov 22 05:11:23 crc kubenswrapper[4818]: I1122 05:11:23.803654 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-ovsdbserver-sb\") pod \"828b46a2-0bd7-430c-8cd1-06be3d81b963\" (UID: \"828b46a2-0bd7-430c-8cd1-06be3d81b963\") " Nov 22 05:11:23 crc kubenswrapper[4818]: I1122 05:11:23.810319 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/828b46a2-0bd7-430c-8cd1-06be3d81b963-kube-api-access-rxcvc" (OuterVolumeSpecName: "kube-api-access-rxcvc") pod "828b46a2-0bd7-430c-8cd1-06be3d81b963" (UID: "828b46a2-0bd7-430c-8cd1-06be3d81b963"). InnerVolumeSpecName "kube-api-access-rxcvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:11:23 crc kubenswrapper[4818]: I1122 05:11:23.865034 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "828b46a2-0bd7-430c-8cd1-06be3d81b963" (UID: "828b46a2-0bd7-430c-8cd1-06be3d81b963"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:11:23 crc kubenswrapper[4818]: I1122 05:11:23.865612 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "828b46a2-0bd7-430c-8cd1-06be3d81b963" (UID: "828b46a2-0bd7-430c-8cd1-06be3d81b963"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:11:23 crc kubenswrapper[4818]: I1122 05:11:23.866319 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-config" (OuterVolumeSpecName: "config") pod "828b46a2-0bd7-430c-8cd1-06be3d81b963" (UID: "828b46a2-0bd7-430c-8cd1-06be3d81b963"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:11:23 crc kubenswrapper[4818]: I1122 05:11:23.866491 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "828b46a2-0bd7-430c-8cd1-06be3d81b963" (UID: "828b46a2-0bd7-430c-8cd1-06be3d81b963"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:11:23 crc kubenswrapper[4818]: I1122 05:11:23.905232 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:23 crc kubenswrapper[4818]: I1122 05:11:23.905273 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:23 crc kubenswrapper[4818]: I1122 05:11:23.905285 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxcvc\" (UniqueName: \"kubernetes.io/projected/828b46a2-0bd7-430c-8cd1-06be3d81b963-kube-api-access-rxcvc\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:23 crc kubenswrapper[4818]: I1122 05:11:23.905293 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:23 crc kubenswrapper[4818]: I1122 05:11:23.905301 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/828b46a2-0bd7-430c-8cd1-06be3d81b963-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:24 crc kubenswrapper[4818]: I1122 05:11:24.137627 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" event={"ID":"828b46a2-0bd7-430c-8cd1-06be3d81b963","Type":"ContainerDied","Data":"bbab8cbc6f7325b55b38ae3989ecccd986b3855b9320754a71a4631d8a6dbba7"} Nov 22 05:11:24 crc kubenswrapper[4818]: I1122 05:11:24.137680 4818 scope.go:117] "RemoveContainer" containerID="a283aefa9ac79d3f17223b04c4310fb68a6c2e743bda42644cccace7cc2d369e" Nov 22 05:11:24 crc kubenswrapper[4818]: I1122 05:11:24.137802 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-745b9ddc8c-cpmmw" Nov 22 05:11:24 crc kubenswrapper[4818]: I1122 05:11:24.176512 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-745b9ddc8c-cpmmw"] Nov 22 05:11:24 crc kubenswrapper[4818]: I1122 05:11:24.182803 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-745b9ddc8c-cpmmw"] Nov 22 05:11:24 crc kubenswrapper[4818]: I1122 05:11:24.304182 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="828b46a2-0bd7-430c-8cd1-06be3d81b963" path="/var/lib/kubelet/pods/828b46a2-0bd7-430c-8cd1-06be3d81b963/volumes" Nov 22 05:11:24 crc kubenswrapper[4818]: I1122 05:11:24.323648 4818 scope.go:117] "RemoveContainer" containerID="c56db2f1c6f7d67c272a1235b8914e89669824b3796414f355ec5eb8bad20cfd" Nov 22 05:11:24 crc kubenswrapper[4818]: E1122 05:11:24.598842 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="0f6203d4-9fff-483c-bc5c-84deb159d47d" Nov 22 05:11:25 crc kubenswrapper[4818]: I1122 05:11:25.151629 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7spb7" event={"ID":"0ca1db80-7d8d-44cf-80bf-afa1c0798182","Type":"ContainerStarted","Data":"fed7bf830da4016586061964fd8e2a8b04bd20aa622442114222ec6e1297420a"} Nov 22 05:11:25 crc kubenswrapper[4818]: I1122 05:11:25.160133 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-n2ldz" event={"ID":"2a8a8454-074a-48c1-b1f2-f800dd6f0f0a","Type":"ContainerStarted","Data":"b54e954468bed2959f1452b89b73005be137fa127ce33fe4114b82bf04b2a03b"} Nov 22 05:11:25 crc kubenswrapper[4818]: I1122 05:11:25.165161 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f6203d4-9fff-483c-bc5c-84deb159d47d","Type":"ContainerStarted","Data":"86b9735df77b1ddcf2089199f40768b0089f2085555bfff643165c859e3a44bc"} Nov 22 05:11:25 crc kubenswrapper[4818]: I1122 05:11:25.165633 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 05:11:25 crc kubenswrapper[4818]: I1122 05:11:25.165630 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0f6203d4-9fff-483c-bc5c-84deb159d47d" containerName="sg-core" containerID="cri-o://7cb97b13bbe6cb54623a632b4f7c9b161afb6c86f35e9e141fd7166dc28a44e2" gracePeriod=30 Nov 22 05:11:25 crc kubenswrapper[4818]: I1122 05:11:25.165699 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0f6203d4-9fff-483c-bc5c-84deb159d47d" containerName="proxy-httpd" containerID="cri-o://86b9735df77b1ddcf2089199f40768b0089f2085555bfff643165c859e3a44bc" gracePeriod=30 Nov 22 05:11:25 crc kubenswrapper[4818]: I1122 05:11:25.165846 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0f6203d4-9fff-483c-bc5c-84deb159d47d" containerName="ceilometer-notification-agent" containerID="cri-o://3c7694200619b732d5d71a5c66213d99c4dc98ba443c3669d13766ec94e389b6" gracePeriod=30 Nov 22 05:11:25 crc kubenswrapper[4818]: I1122 05:11:25.177519 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522"} Nov 22 05:11:25 crc kubenswrapper[4818]: I1122 05:11:25.179881 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-7spb7" podStartSLOduration=2.075687608 podStartE2EDuration="42.179860278s" podCreationTimestamp="2025-11-22 05:10:43 +0000 UTC" firstStartedPulling="2025-11-22 05:10:44.254451593 +0000 UTC m=+1396.828868120" lastFinishedPulling="2025-11-22 05:11:24.358624263 +0000 UTC m=+1436.933040790" observedRunningTime="2025-11-22 05:11:25.170607924 +0000 UTC m=+1437.745024481" watchObservedRunningTime="2025-11-22 05:11:25.179860278 +0000 UTC m=+1437.754276835" Nov 22 05:11:25 crc kubenswrapper[4818]: I1122 05:11:25.182129 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7b464767cb-jcwc8" event={"ID":"4e48ac54-5e91-4747-b9e9-28aeb21180fd","Type":"ContainerStarted","Data":"9123816d2694b90a3e6bfc6e2880ea54bea44b18c9946d307a6c450f9a6c9fc8"} Nov 22 05:11:25 crc kubenswrapper[4818]: I1122 05:11:25.182608 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:25 crc kubenswrapper[4818]: I1122 05:11:25.182779 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:25 crc kubenswrapper[4818]: I1122 05:11:25.225781 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-n2ldz" podStartSLOduration=2.446133724 podStartE2EDuration="42.225761729s" podCreationTimestamp="2025-11-22 05:10:43 +0000 UTC" firstStartedPulling="2025-11-22 05:10:44.580040226 +0000 UTC m=+1397.154456753" lastFinishedPulling="2025-11-22 05:11:24.359668231 +0000 UTC m=+1436.934084758" observedRunningTime="2025-11-22 05:11:25.197144903 +0000 UTC m=+1437.771561450" watchObservedRunningTime="2025-11-22 05:11:25.225761729 +0000 UTC m=+1437.800178256" Nov 22 05:11:25 crc kubenswrapper[4818]: I1122 05:11:25.250791 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7b464767cb-jcwc8" podStartSLOduration=7.250769737 podStartE2EDuration="7.250769737s" podCreationTimestamp="2025-11-22 05:11:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:11:25.244848771 +0000 UTC m=+1437.819265338" watchObservedRunningTime="2025-11-22 05:11:25.250769737 +0000 UTC m=+1437.825186264" Nov 22 05:11:26 crc kubenswrapper[4818]: I1122 05:11:26.226687 4818 generic.go:334] "Generic (PLEG): container finished" podID="0f6203d4-9fff-483c-bc5c-84deb159d47d" containerID="86b9735df77b1ddcf2089199f40768b0089f2085555bfff643165c859e3a44bc" exitCode=0 Nov 22 05:11:26 crc kubenswrapper[4818]: I1122 05:11:26.227671 4818 generic.go:334] "Generic (PLEG): container finished" podID="0f6203d4-9fff-483c-bc5c-84deb159d47d" containerID="7cb97b13bbe6cb54623a632b4f7c9b161afb6c86f35e9e141fd7166dc28a44e2" exitCode=2 Nov 22 05:11:26 crc kubenswrapper[4818]: I1122 05:11:26.226769 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f6203d4-9fff-483c-bc5c-84deb159d47d","Type":"ContainerDied","Data":"86b9735df77b1ddcf2089199f40768b0089f2085555bfff643165c859e3a44bc"} Nov 22 05:11:26 crc kubenswrapper[4818]: I1122 05:11:26.227798 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f6203d4-9fff-483c-bc5c-84deb159d47d","Type":"ContainerDied","Data":"7cb97b13bbe6cb54623a632b4f7c9b161afb6c86f35e9e141fd7166dc28a44e2"} Nov 22 05:11:26 crc kubenswrapper[4818]: I1122 05:11:26.230166 4818 generic.go:334] "Generic (PLEG): container finished" podID="f84522a7-f356-40c9-8ea0-13b1503993fe" containerID="7d8c16c52256a7f940166ce1567aa8686f167a0d8de2b989b4302e9a7cb4adf9" exitCode=0 Nov 22 05:11:26 crc kubenswrapper[4818]: I1122 05:11:26.230222 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-rd5pw" event={"ID":"f84522a7-f356-40c9-8ea0-13b1503993fe","Type":"ContainerDied","Data":"7d8c16c52256a7f940166ce1567aa8686f167a0d8de2b989b4302e9a7cb4adf9"} Nov 22 05:11:26 crc kubenswrapper[4818]: I1122 05:11:26.439351 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:27 crc kubenswrapper[4818]: I1122 05:11:27.613521 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-rd5pw" Nov 22 05:11:27 crc kubenswrapper[4818]: I1122 05:11:27.772162 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f84522a7-f356-40c9-8ea0-13b1503993fe-config\") pod \"f84522a7-f356-40c9-8ea0-13b1503993fe\" (UID: \"f84522a7-f356-40c9-8ea0-13b1503993fe\") " Nov 22 05:11:27 crc kubenswrapper[4818]: I1122 05:11:27.772219 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njwqs\" (UniqueName: \"kubernetes.io/projected/f84522a7-f356-40c9-8ea0-13b1503993fe-kube-api-access-njwqs\") pod \"f84522a7-f356-40c9-8ea0-13b1503993fe\" (UID: \"f84522a7-f356-40c9-8ea0-13b1503993fe\") " Nov 22 05:11:27 crc kubenswrapper[4818]: I1122 05:11:27.772337 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f84522a7-f356-40c9-8ea0-13b1503993fe-combined-ca-bundle\") pod \"f84522a7-f356-40c9-8ea0-13b1503993fe\" (UID: \"f84522a7-f356-40c9-8ea0-13b1503993fe\") " Nov 22 05:11:27 crc kubenswrapper[4818]: I1122 05:11:27.783142 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f84522a7-f356-40c9-8ea0-13b1503993fe-kube-api-access-njwqs" (OuterVolumeSpecName: "kube-api-access-njwqs") pod "f84522a7-f356-40c9-8ea0-13b1503993fe" (UID: "f84522a7-f356-40c9-8ea0-13b1503993fe"). InnerVolumeSpecName "kube-api-access-njwqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:11:27 crc kubenswrapper[4818]: I1122 05:11:27.806933 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f84522a7-f356-40c9-8ea0-13b1503993fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f84522a7-f356-40c9-8ea0-13b1503993fe" (UID: "f84522a7-f356-40c9-8ea0-13b1503993fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:27 crc kubenswrapper[4818]: I1122 05:11:27.837803 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f84522a7-f356-40c9-8ea0-13b1503993fe-config" (OuterVolumeSpecName: "config") pod "f84522a7-f356-40c9-8ea0-13b1503993fe" (UID: "f84522a7-f356-40c9-8ea0-13b1503993fe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:27 crc kubenswrapper[4818]: I1122 05:11:27.874403 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/f84522a7-f356-40c9-8ea0-13b1503993fe-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:27 crc kubenswrapper[4818]: I1122 05:11:27.874621 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njwqs\" (UniqueName: \"kubernetes.io/projected/f84522a7-f356-40c9-8ea0-13b1503993fe-kube-api-access-njwqs\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:27 crc kubenswrapper[4818]: I1122 05:11:27.874888 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f84522a7-f356-40c9-8ea0-13b1503993fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.254755 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-rd5pw" event={"ID":"f84522a7-f356-40c9-8ea0-13b1503993fe","Type":"ContainerDied","Data":"ba1eb2580d8e6f81a2ce1117a4ff385f4c956bd774860a8186f6bf80af7d14e7"} Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.254796 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba1eb2580d8e6f81a2ce1117a4ff385f4c956bd774860a8186f6bf80af7d14e7" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.254892 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-rd5pw" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.505247 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-ts7sf"] Nov 22 05:11:28 crc kubenswrapper[4818]: E1122 05:11:28.505686 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="828b46a2-0bd7-430c-8cd1-06be3d81b963" containerName="dnsmasq-dns" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.505711 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="828b46a2-0bd7-430c-8cd1-06be3d81b963" containerName="dnsmasq-dns" Nov 22 05:11:28 crc kubenswrapper[4818]: E1122 05:11:28.505743 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="828b46a2-0bd7-430c-8cd1-06be3d81b963" containerName="init" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.505752 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="828b46a2-0bd7-430c-8cd1-06be3d81b963" containerName="init" Nov 22 05:11:28 crc kubenswrapper[4818]: E1122 05:11:28.505768 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f84522a7-f356-40c9-8ea0-13b1503993fe" containerName="neutron-db-sync" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.505776 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f84522a7-f356-40c9-8ea0-13b1503993fe" containerName="neutron-db-sync" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.505975 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f84522a7-f356-40c9-8ea0-13b1503993fe" containerName="neutron-db-sync" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.505994 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="828b46a2-0bd7-430c-8cd1-06be3d81b963" containerName="dnsmasq-dns" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.506836 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.538438 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-ts7sf"] Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.635733 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-85b5b4f6cd-lp7b9"] Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.637094 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.639418 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-jn2fl" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.639654 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.639794 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.641103 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.653168 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-85b5b4f6cd-lp7b9"] Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.689502 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-config\") pod \"dnsmasq-dns-7b946d459c-ts7sf\" (UID: \"8e82d7f2-2913-48be-ae98-28596edf9d4a\") " pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.689566 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-ovsdbserver-nb\") pod \"dnsmasq-dns-7b946d459c-ts7sf\" (UID: \"8e82d7f2-2913-48be-ae98-28596edf9d4a\") " pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.689699 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-ovsdbserver-sb\") pod \"dnsmasq-dns-7b946d459c-ts7sf\" (UID: \"8e82d7f2-2913-48be-ae98-28596edf9d4a\") " pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.689909 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9s77\" (UniqueName: \"kubernetes.io/projected/8e82d7f2-2913-48be-ae98-28596edf9d4a-kube-api-access-q9s77\") pod \"dnsmasq-dns-7b946d459c-ts7sf\" (UID: \"8e82d7f2-2913-48be-ae98-28596edf9d4a\") " pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.690030 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-dns-svc\") pod \"dnsmasq-dns-7b946d459c-ts7sf\" (UID: \"8e82d7f2-2913-48be-ae98-28596edf9d4a\") " pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.791248 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvx4z\" (UniqueName: \"kubernetes.io/projected/5578cc92-9b08-4c31-a73f-9fcda619ee85-kube-api-access-jvx4z\") pod \"neutron-85b5b4f6cd-lp7b9\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.791538 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-config\") pod \"dnsmasq-dns-7b946d459c-ts7sf\" (UID: \"8e82d7f2-2913-48be-ae98-28596edf9d4a\") " pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.791576 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-ovsdbserver-nb\") pod \"dnsmasq-dns-7b946d459c-ts7sf\" (UID: \"8e82d7f2-2913-48be-ae98-28596edf9d4a\") " pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.791607 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-ovsdbserver-sb\") pod \"dnsmasq-dns-7b946d459c-ts7sf\" (UID: \"8e82d7f2-2913-48be-ae98-28596edf9d4a\") " pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.791645 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-ovndb-tls-certs\") pod \"neutron-85b5b4f6cd-lp7b9\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.791661 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-combined-ca-bundle\") pod \"neutron-85b5b4f6cd-lp7b9\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.791843 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9s77\" (UniqueName: \"kubernetes.io/projected/8e82d7f2-2913-48be-ae98-28596edf9d4a-kube-api-access-q9s77\") pod \"dnsmasq-dns-7b946d459c-ts7sf\" (UID: \"8e82d7f2-2913-48be-ae98-28596edf9d4a\") " pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.791969 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-config\") pod \"neutron-85b5b4f6cd-lp7b9\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.792028 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-dns-svc\") pod \"dnsmasq-dns-7b946d459c-ts7sf\" (UID: \"8e82d7f2-2913-48be-ae98-28596edf9d4a\") " pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.792044 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-httpd-config\") pod \"neutron-85b5b4f6cd-lp7b9\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.792521 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-ovsdbserver-nb\") pod \"dnsmasq-dns-7b946d459c-ts7sf\" (UID: \"8e82d7f2-2913-48be-ae98-28596edf9d4a\") " pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.792865 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-ovsdbserver-sb\") pod \"dnsmasq-dns-7b946d459c-ts7sf\" (UID: \"8e82d7f2-2913-48be-ae98-28596edf9d4a\") " pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.792955 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-config\") pod \"dnsmasq-dns-7b946d459c-ts7sf\" (UID: \"8e82d7f2-2913-48be-ae98-28596edf9d4a\") " pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.793011 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-dns-svc\") pod \"dnsmasq-dns-7b946d459c-ts7sf\" (UID: \"8e82d7f2-2913-48be-ae98-28596edf9d4a\") " pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.808986 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9s77\" (UniqueName: \"kubernetes.io/projected/8e82d7f2-2913-48be-ae98-28596edf9d4a-kube-api-access-q9s77\") pod \"dnsmasq-dns-7b946d459c-ts7sf\" (UID: \"8e82d7f2-2913-48be-ae98-28596edf9d4a\") " pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.826712 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.894214 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-httpd-config\") pod \"neutron-85b5b4f6cd-lp7b9\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.894297 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvx4z\" (UniqueName: \"kubernetes.io/projected/5578cc92-9b08-4c31-a73f-9fcda619ee85-kube-api-access-jvx4z\") pod \"neutron-85b5b4f6cd-lp7b9\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.894371 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-combined-ca-bundle\") pod \"neutron-85b5b4f6cd-lp7b9\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.894389 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-ovndb-tls-certs\") pod \"neutron-85b5b4f6cd-lp7b9\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.894452 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-config\") pod \"neutron-85b5b4f6cd-lp7b9\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.901374 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-ovndb-tls-certs\") pod \"neutron-85b5b4f6cd-lp7b9\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.902491 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-httpd-config\") pod \"neutron-85b5b4f6cd-lp7b9\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.906678 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-config\") pod \"neutron-85b5b4f6cd-lp7b9\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.908825 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-combined-ca-bundle\") pod \"neutron-85b5b4f6cd-lp7b9\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.914338 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvx4z\" (UniqueName: \"kubernetes.io/projected/5578cc92-9b08-4c31-a73f-9fcda619ee85-kube-api-access-jvx4z\") pod \"neutron-85b5b4f6cd-lp7b9\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:11:28 crc kubenswrapper[4818]: I1122 05:11:28.955711 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:11:29 crc kubenswrapper[4818]: I1122 05:11:29.288527 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-ts7sf"] Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.281433 4818 generic.go:334] "Generic (PLEG): container finished" podID="0f6203d4-9fff-483c-bc5c-84deb159d47d" containerID="3c7694200619b732d5d71a5c66213d99c4dc98ba443c3669d13766ec94e389b6" exitCode=0 Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.281490 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f6203d4-9fff-483c-bc5c-84deb159d47d","Type":"ContainerDied","Data":"3c7694200619b732d5d71a5c66213d99c4dc98ba443c3669d13766ec94e389b6"} Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.282046 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f6203d4-9fff-483c-bc5c-84deb159d47d","Type":"ContainerDied","Data":"bfe8c2dc1c03792e20e9966cc8989dd047e089919fb0d47ee582099b9a40c8da"} Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.282061 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfe8c2dc1c03792e20e9966cc8989dd047e089919fb0d47ee582099b9a40c8da" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.283467 4818 generic.go:334] "Generic (PLEG): container finished" podID="8e82d7f2-2913-48be-ae98-28596edf9d4a" containerID="fa16ed87ba70a91bb87f1391b24f651455dc50a14bf20b9b1de2ca2f53026cb9" exitCode=0 Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.283501 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" event={"ID":"8e82d7f2-2913-48be-ae98-28596edf9d4a","Type":"ContainerDied","Data":"fa16ed87ba70a91bb87f1391b24f651455dc50a14bf20b9b1de2ca2f53026cb9"} Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.283518 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" event={"ID":"8e82d7f2-2913-48be-ae98-28596edf9d4a","Type":"ContainerStarted","Data":"27072eaa1ea39a5268e2f28a3ca4257df6fb79d3a809abded046fab16f024099"} Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.382854 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.522158 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-sg-core-conf-yaml\") pod \"0f6203d4-9fff-483c-bc5c-84deb159d47d\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.522660 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-config-data\") pod \"0f6203d4-9fff-483c-bc5c-84deb159d47d\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.522702 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-combined-ca-bundle\") pod \"0f6203d4-9fff-483c-bc5c-84deb159d47d\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.522781 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f6203d4-9fff-483c-bc5c-84deb159d47d-run-httpd\") pod \"0f6203d4-9fff-483c-bc5c-84deb159d47d\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.522825 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f6203d4-9fff-483c-bc5c-84deb159d47d-log-httpd\") pod \"0f6203d4-9fff-483c-bc5c-84deb159d47d\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.522886 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x826k\" (UniqueName: \"kubernetes.io/projected/0f6203d4-9fff-483c-bc5c-84deb159d47d-kube-api-access-x826k\") pod \"0f6203d4-9fff-483c-bc5c-84deb159d47d\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.522990 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-scripts\") pod \"0f6203d4-9fff-483c-bc5c-84deb159d47d\" (UID: \"0f6203d4-9fff-483c-bc5c-84deb159d47d\") " Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.523271 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f6203d4-9fff-483c-bc5c-84deb159d47d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0f6203d4-9fff-483c-bc5c-84deb159d47d" (UID: "0f6203d4-9fff-483c-bc5c-84deb159d47d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.523609 4818 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f6203d4-9fff-483c-bc5c-84deb159d47d-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.523788 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f6203d4-9fff-483c-bc5c-84deb159d47d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0f6203d4-9fff-483c-bc5c-84deb159d47d" (UID: "0f6203d4-9fff-483c-bc5c-84deb159d47d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.528550 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-scripts" (OuterVolumeSpecName: "scripts") pod "0f6203d4-9fff-483c-bc5c-84deb159d47d" (UID: "0f6203d4-9fff-483c-bc5c-84deb159d47d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.528838 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f6203d4-9fff-483c-bc5c-84deb159d47d-kube-api-access-x826k" (OuterVolumeSpecName: "kube-api-access-x826k") pod "0f6203d4-9fff-483c-bc5c-84deb159d47d" (UID: "0f6203d4-9fff-483c-bc5c-84deb159d47d"). InnerVolumeSpecName "kube-api-access-x826k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.551517 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0f6203d4-9fff-483c-bc5c-84deb159d47d" (UID: "0f6203d4-9fff-483c-bc5c-84deb159d47d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.575203 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f6203d4-9fff-483c-bc5c-84deb159d47d" (UID: "0f6203d4-9fff-483c-bc5c-84deb159d47d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.603128 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-config-data" (OuterVolumeSpecName: "config-data") pod "0f6203d4-9fff-483c-bc5c-84deb159d47d" (UID: "0f6203d4-9fff-483c-bc5c-84deb159d47d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.627224 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.627323 4818 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.627338 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.627346 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f6203d4-9fff-483c-bc5c-84deb159d47d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.627355 4818 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f6203d4-9fff-483c-bc5c-84deb159d47d-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.627363 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x826k\" (UniqueName: \"kubernetes.io/projected/0f6203d4-9fff-483c-bc5c-84deb159d47d-kube-api-access-x826k\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.646907 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-64754b546f-kqpj9"] Nov 22 05:11:30 crc kubenswrapper[4818]: E1122 05:11:30.647314 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f6203d4-9fff-483c-bc5c-84deb159d47d" containerName="sg-core" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.647336 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f6203d4-9fff-483c-bc5c-84deb159d47d" containerName="sg-core" Nov 22 05:11:30 crc kubenswrapper[4818]: E1122 05:11:30.647373 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f6203d4-9fff-483c-bc5c-84deb159d47d" containerName="ceilometer-notification-agent" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.647382 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f6203d4-9fff-483c-bc5c-84deb159d47d" containerName="ceilometer-notification-agent" Nov 22 05:11:30 crc kubenswrapper[4818]: E1122 05:11:30.647397 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f6203d4-9fff-483c-bc5c-84deb159d47d" containerName="proxy-httpd" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.647406 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f6203d4-9fff-483c-bc5c-84deb159d47d" containerName="proxy-httpd" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.647617 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f6203d4-9fff-483c-bc5c-84deb159d47d" containerName="sg-core" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.647639 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f6203d4-9fff-483c-bc5c-84deb159d47d" containerName="proxy-httpd" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.647660 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f6203d4-9fff-483c-bc5c-84deb159d47d" containerName="ceilometer-notification-agent" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.648655 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.650575 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.651178 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.665738 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-64754b546f-kqpj9"] Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.729371 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-internal-tls-certs\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.729454 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-httpd-config\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.729537 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-config\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.729569 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9r8m\" (UniqueName: \"kubernetes.io/projected/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-kube-api-access-p9r8m\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.729618 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-combined-ca-bundle\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.729666 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-ovndb-tls-certs\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.729695 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-public-tls-certs\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.813519 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-85b5b4f6cd-lp7b9"] Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.831094 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-config\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.831284 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9r8m\" (UniqueName: \"kubernetes.io/projected/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-kube-api-access-p9r8m\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.831437 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-combined-ca-bundle\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.831555 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-ovndb-tls-certs\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.831662 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-public-tls-certs\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.831796 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-internal-tls-certs\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.831918 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-httpd-config\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.835727 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-httpd-config\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.837447 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-public-tls-certs\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.837872 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-ovndb-tls-certs\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.838428 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-config\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.840766 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-internal-tls-certs\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.845837 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-combined-ca-bundle\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.852626 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9r8m\" (UniqueName: \"kubernetes.io/projected/81f8a1cb-f2b5-429b-b12b-d2ff49b2e888-kube-api-access-p9r8m\") pod \"neutron-64754b546f-kqpj9\" (UID: \"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888\") " pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:30 crc kubenswrapper[4818]: I1122 05:11:30.989773 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.331237 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" event={"ID":"8e82d7f2-2913-48be-ae98-28596edf9d4a","Type":"ContainerStarted","Data":"1307d498f0f33be235accb53597a8787d249fab7e5cc31b0c95f406288ac414e"} Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.332737 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.335376 4818 generic.go:334] "Generic (PLEG): container finished" podID="2a8a8454-074a-48c1-b1f2-f800dd6f0f0a" containerID="b54e954468bed2959f1452b89b73005be137fa127ce33fe4114b82bf04b2a03b" exitCode=0 Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.335691 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-n2ldz" event={"ID":"2a8a8454-074a-48c1-b1f2-f800dd6f0f0a","Type":"ContainerDied","Data":"b54e954468bed2959f1452b89b73005be137fa127ce33fe4114b82bf04b2a03b"} Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.339015 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.339041 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85b5b4f6cd-lp7b9" event={"ID":"5578cc92-9b08-4c31-a73f-9fcda619ee85","Type":"ContainerStarted","Data":"9d4df4d517e61088fca1643c98ccaae797e79aa61382a96b1d80a483e5add0d2"} Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.339090 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85b5b4f6cd-lp7b9" event={"ID":"5578cc92-9b08-4c31-a73f-9fcda619ee85","Type":"ContainerStarted","Data":"8331b76f4dea87944ae11fc2b5fcb21a8047259b56bd8d8fc8d6e021f6c31446"} Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.339109 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85b5b4f6cd-lp7b9" event={"ID":"5578cc92-9b08-4c31-a73f-9fcda619ee85","Type":"ContainerStarted","Data":"06e597820204a4d2bf0d67108be5daf4600ea2278275ec786036526dfd5c64e1"} Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.350865 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" podStartSLOduration=3.350843535 podStartE2EDuration="3.350843535s" podCreationTimestamp="2025-11-22 05:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:11:31.350557307 +0000 UTC m=+1443.924973904" watchObservedRunningTime="2025-11-22 05:11:31.350843535 +0000 UTC m=+1443.925260072" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.391029 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-85b5b4f6cd-lp7b9" podStartSLOduration=3.391009684 podStartE2EDuration="3.391009684s" podCreationTimestamp="2025-11-22 05:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:11:31.377473417 +0000 UTC m=+1443.951889964" watchObservedRunningTime="2025-11-22 05:11:31.391009684 +0000 UTC m=+1443.965426211" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.489753 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.501539 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.523477 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.525546 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.529077 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.529100 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.547178 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.594238 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-64754b546f-kqpj9"] Nov 22 05:11:31 crc kubenswrapper[4818]: W1122 05:11:31.600610 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81f8a1cb_f2b5_429b_b12b_d2ff49b2e888.slice/crio-dd3e1582dfe388770aa491d7d36847157bcbf9321deadaee388531808b443f35 WatchSource:0}: Error finding container dd3e1582dfe388770aa491d7d36847157bcbf9321deadaee388531808b443f35: Status 404 returned error can't find the container with id dd3e1582dfe388770aa491d7d36847157bcbf9321deadaee388531808b443f35 Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.655452 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.655509 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.655543 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-run-httpd\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.655609 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-config-data\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.655662 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-scripts\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.655678 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-log-httpd\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.655706 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvsjb\" (UniqueName: \"kubernetes.io/projected/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-kube-api-access-vvsjb\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.757347 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.757393 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.757424 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-run-httpd\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.757449 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-config-data\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.757488 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-scripts\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.757505 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-log-httpd\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.757535 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvsjb\" (UniqueName: \"kubernetes.io/projected/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-kube-api-access-vvsjb\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.758154 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-run-httpd\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.758149 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-log-httpd\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.762104 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-scripts\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.763279 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-config-data\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.765892 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.766785 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.779644 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvsjb\" (UniqueName: \"kubernetes.io/projected/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-kube-api-access-vvsjb\") pod \"ceilometer-0\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " pod="openstack/ceilometer-0" Nov 22 05:11:31 crc kubenswrapper[4818]: I1122 05:11:31.863620 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:11:32 crc kubenswrapper[4818]: I1122 05:11:32.305896 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f6203d4-9fff-483c-bc5c-84deb159d47d" path="/var/lib/kubelet/pods/0f6203d4-9fff-483c-bc5c-84deb159d47d/volumes" Nov 22 05:11:32 crc kubenswrapper[4818]: I1122 05:11:32.324718 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:11:32 crc kubenswrapper[4818]: I1122 05:11:32.346686 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64754b546f-kqpj9" event={"ID":"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888","Type":"ContainerStarted","Data":"7b601fdd2aa38b69bdd71a4ab025b55c20439198ae414d7742a5f59fa8c7521f"} Nov 22 05:11:32 crc kubenswrapper[4818]: I1122 05:11:32.346741 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64754b546f-kqpj9" event={"ID":"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888","Type":"ContainerStarted","Data":"42e7f858c43c1d87003ec682c891b6ad63fb3344bc7057a116a1e81bee7e699b"} Nov 22 05:11:32 crc kubenswrapper[4818]: I1122 05:11:32.346760 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64754b546f-kqpj9" event={"ID":"81f8a1cb-f2b5-429b-b12b-d2ff49b2e888","Type":"ContainerStarted","Data":"dd3e1582dfe388770aa491d7d36847157bcbf9321deadaee388531808b443f35"} Nov 22 05:11:32 crc kubenswrapper[4818]: I1122 05:11:32.348076 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c105a582-4a5b-4c1b-aac3-bbde151fcb1b","Type":"ContainerStarted","Data":"913bea3087c6389e694663ed264cd7da7ca3c53ba4fbd657f836e3897e944fc5"} Nov 22 05:11:32 crc kubenswrapper[4818]: I1122 05:11:32.348338 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:11:32 crc kubenswrapper[4818]: I1122 05:11:32.632940 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-n2ldz" Nov 22 05:11:32 crc kubenswrapper[4818]: I1122 05:11:32.775795 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2a8a8454-074a-48c1-b1f2-f800dd6f0f0a-db-sync-config-data\") pod \"2a8a8454-074a-48c1-b1f2-f800dd6f0f0a\" (UID: \"2a8a8454-074a-48c1-b1f2-f800dd6f0f0a\") " Nov 22 05:11:32 crc kubenswrapper[4818]: I1122 05:11:32.776049 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-577sx\" (UniqueName: \"kubernetes.io/projected/2a8a8454-074a-48c1-b1f2-f800dd6f0f0a-kube-api-access-577sx\") pod \"2a8a8454-074a-48c1-b1f2-f800dd6f0f0a\" (UID: \"2a8a8454-074a-48c1-b1f2-f800dd6f0f0a\") " Nov 22 05:11:32 crc kubenswrapper[4818]: I1122 05:11:32.776118 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a8a8454-074a-48c1-b1f2-f800dd6f0f0a-combined-ca-bundle\") pod \"2a8a8454-074a-48c1-b1f2-f800dd6f0f0a\" (UID: \"2a8a8454-074a-48c1-b1f2-f800dd6f0f0a\") " Nov 22 05:11:32 crc kubenswrapper[4818]: I1122 05:11:32.782424 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a8a8454-074a-48c1-b1f2-f800dd6f0f0a-kube-api-access-577sx" (OuterVolumeSpecName: "kube-api-access-577sx") pod "2a8a8454-074a-48c1-b1f2-f800dd6f0f0a" (UID: "2a8a8454-074a-48c1-b1f2-f800dd6f0f0a"). InnerVolumeSpecName "kube-api-access-577sx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:11:32 crc kubenswrapper[4818]: I1122 05:11:32.782492 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a8a8454-074a-48c1-b1f2-f800dd6f0f0a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2a8a8454-074a-48c1-b1f2-f800dd6f0f0a" (UID: "2a8a8454-074a-48c1-b1f2-f800dd6f0f0a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:32 crc kubenswrapper[4818]: I1122 05:11:32.827058 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a8a8454-074a-48c1-b1f2-f800dd6f0f0a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a8a8454-074a-48c1-b1f2-f800dd6f0f0a" (UID: "2a8a8454-074a-48c1-b1f2-f800dd6f0f0a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:32 crc kubenswrapper[4818]: I1122 05:11:32.878754 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-577sx\" (UniqueName: \"kubernetes.io/projected/2a8a8454-074a-48c1-b1f2-f800dd6f0f0a-kube-api-access-577sx\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:32 crc kubenswrapper[4818]: I1122 05:11:32.878787 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a8a8454-074a-48c1-b1f2-f800dd6f0f0a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:32 crc kubenswrapper[4818]: I1122 05:11:32.878796 4818 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2a8a8454-074a-48c1-b1f2-f800dd6f0f0a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.358850 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-n2ldz" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.364375 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-n2ldz" event={"ID":"2a8a8454-074a-48c1-b1f2-f800dd6f0f0a","Type":"ContainerDied","Data":"2874f707c5ab53823774fd6a42fca4daf3e1b8cd08ce7f7d50f19bdfe3ddf2ec"} Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.364939 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2874f707c5ab53823774fd6a42fca4daf3e1b8cd08ce7f7d50f19bdfe3ddf2ec" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.365024 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.391032 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-64754b546f-kqpj9" podStartSLOduration=3.39101594 podStartE2EDuration="3.39101594s" podCreationTimestamp="2025-11-22 05:11:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:11:33.390348423 +0000 UTC m=+1445.964764950" watchObservedRunningTime="2025-11-22 05:11:33.39101594 +0000 UTC m=+1445.965432467" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.675747 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-fb9947b9-nbdwz"] Nov 22 05:11:33 crc kubenswrapper[4818]: E1122 05:11:33.676217 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a8a8454-074a-48c1-b1f2-f800dd6f0f0a" containerName="barbican-db-sync" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.676240 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a8a8454-074a-48c1-b1f2-f800dd6f0f0a" containerName="barbican-db-sync" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.676460 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a8a8454-074a-48c1-b1f2-f800dd6f0f0a" containerName="barbican-db-sync" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.677599 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-fb9947b9-nbdwz" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.683618 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-j9jxn" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.683794 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.688208 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.700934 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5c65499f7d-s6w7z"] Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.704097 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.706577 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.711264 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-fb9947b9-nbdwz"] Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.765345 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5c65499f7d-s6w7z"] Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.796268 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5925d65-97cd-493a-8114-6ed07de4386d-logs\") pod \"barbican-keystone-listener-5c65499f7d-s6w7z\" (UID: \"a5925d65-97cd-493a-8114-6ed07de4386d\") " pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.796315 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3958bab3-ce66-4d2d-8a76-75e1b7593199-config-data-custom\") pod \"barbican-worker-fb9947b9-nbdwz\" (UID: \"3958bab3-ce66-4d2d-8a76-75e1b7593199\") " pod="openstack/barbican-worker-fb9947b9-nbdwz" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.796342 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5925d65-97cd-493a-8114-6ed07de4386d-config-data\") pod \"barbican-keystone-listener-5c65499f7d-s6w7z\" (UID: \"a5925d65-97cd-493a-8114-6ed07de4386d\") " pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.796375 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mz625\" (UniqueName: \"kubernetes.io/projected/a5925d65-97cd-493a-8114-6ed07de4386d-kube-api-access-mz625\") pod \"barbican-keystone-listener-5c65499f7d-s6w7z\" (UID: \"a5925d65-97cd-493a-8114-6ed07de4386d\") " pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.796398 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5925d65-97cd-493a-8114-6ed07de4386d-combined-ca-bundle\") pod \"barbican-keystone-listener-5c65499f7d-s6w7z\" (UID: \"a5925d65-97cd-493a-8114-6ed07de4386d\") " pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.796416 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5925d65-97cd-493a-8114-6ed07de4386d-config-data-custom\") pod \"barbican-keystone-listener-5c65499f7d-s6w7z\" (UID: \"a5925d65-97cd-493a-8114-6ed07de4386d\") " pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.796542 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3958bab3-ce66-4d2d-8a76-75e1b7593199-combined-ca-bundle\") pod \"barbican-worker-fb9947b9-nbdwz\" (UID: \"3958bab3-ce66-4d2d-8a76-75e1b7593199\") " pod="openstack/barbican-worker-fb9947b9-nbdwz" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.796630 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3958bab3-ce66-4d2d-8a76-75e1b7593199-logs\") pod \"barbican-worker-fb9947b9-nbdwz\" (UID: \"3958bab3-ce66-4d2d-8a76-75e1b7593199\") " pod="openstack/barbican-worker-fb9947b9-nbdwz" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.796825 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3958bab3-ce66-4d2d-8a76-75e1b7593199-config-data\") pod \"barbican-worker-fb9947b9-nbdwz\" (UID: \"3958bab3-ce66-4d2d-8a76-75e1b7593199\") " pod="openstack/barbican-worker-fb9947b9-nbdwz" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.796894 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7k89\" (UniqueName: \"kubernetes.io/projected/3958bab3-ce66-4d2d-8a76-75e1b7593199-kube-api-access-b7k89\") pod \"barbican-worker-fb9947b9-nbdwz\" (UID: \"3958bab3-ce66-4d2d-8a76-75e1b7593199\") " pod="openstack/barbican-worker-fb9947b9-nbdwz" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.870166 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-ts7sf"] Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.883207 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-rrhr2"] Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.886066 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.901877 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5925d65-97cd-493a-8114-6ed07de4386d-logs\") pod \"barbican-keystone-listener-5c65499f7d-s6w7z\" (UID: \"a5925d65-97cd-493a-8114-6ed07de4386d\") " pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.901918 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3958bab3-ce66-4d2d-8a76-75e1b7593199-config-data-custom\") pod \"barbican-worker-fb9947b9-nbdwz\" (UID: \"3958bab3-ce66-4d2d-8a76-75e1b7593199\") " pod="openstack/barbican-worker-fb9947b9-nbdwz" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.901937 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5925d65-97cd-493a-8114-6ed07de4386d-config-data\") pod \"barbican-keystone-listener-5c65499f7d-s6w7z\" (UID: \"a5925d65-97cd-493a-8114-6ed07de4386d\") " pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.901967 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mz625\" (UniqueName: \"kubernetes.io/projected/a5925d65-97cd-493a-8114-6ed07de4386d-kube-api-access-mz625\") pod \"barbican-keystone-listener-5c65499f7d-s6w7z\" (UID: \"a5925d65-97cd-493a-8114-6ed07de4386d\") " pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.901988 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5925d65-97cd-493a-8114-6ed07de4386d-combined-ca-bundle\") pod \"barbican-keystone-listener-5c65499f7d-s6w7z\" (UID: \"a5925d65-97cd-493a-8114-6ed07de4386d\") " pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.902003 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5925d65-97cd-493a-8114-6ed07de4386d-config-data-custom\") pod \"barbican-keystone-listener-5c65499f7d-s6w7z\" (UID: \"a5925d65-97cd-493a-8114-6ed07de4386d\") " pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.902033 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3958bab3-ce66-4d2d-8a76-75e1b7593199-combined-ca-bundle\") pod \"barbican-worker-fb9947b9-nbdwz\" (UID: \"3958bab3-ce66-4d2d-8a76-75e1b7593199\") " pod="openstack/barbican-worker-fb9947b9-nbdwz" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.902059 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3958bab3-ce66-4d2d-8a76-75e1b7593199-logs\") pod \"barbican-worker-fb9947b9-nbdwz\" (UID: \"3958bab3-ce66-4d2d-8a76-75e1b7593199\") " pod="openstack/barbican-worker-fb9947b9-nbdwz" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.902087 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3958bab3-ce66-4d2d-8a76-75e1b7593199-config-data\") pod \"barbican-worker-fb9947b9-nbdwz\" (UID: \"3958bab3-ce66-4d2d-8a76-75e1b7593199\") " pod="openstack/barbican-worker-fb9947b9-nbdwz" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.902108 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7k89\" (UniqueName: \"kubernetes.io/projected/3958bab3-ce66-4d2d-8a76-75e1b7593199-kube-api-access-b7k89\") pod \"barbican-worker-fb9947b9-nbdwz\" (UID: \"3958bab3-ce66-4d2d-8a76-75e1b7593199\") " pod="openstack/barbican-worker-fb9947b9-nbdwz" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.925181 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5925d65-97cd-493a-8114-6ed07de4386d-logs\") pod \"barbican-keystone-listener-5c65499f7d-s6w7z\" (UID: \"a5925d65-97cd-493a-8114-6ed07de4386d\") " pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.926860 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-rrhr2"] Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.929607 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3958bab3-ce66-4d2d-8a76-75e1b7593199-logs\") pod \"barbican-worker-fb9947b9-nbdwz\" (UID: \"3958bab3-ce66-4d2d-8a76-75e1b7593199\") " pod="openstack/barbican-worker-fb9947b9-nbdwz" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.933122 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3958bab3-ce66-4d2d-8a76-75e1b7593199-config-data-custom\") pod \"barbican-worker-fb9947b9-nbdwz\" (UID: \"3958bab3-ce66-4d2d-8a76-75e1b7593199\") " pod="openstack/barbican-worker-fb9947b9-nbdwz" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.936738 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3958bab3-ce66-4d2d-8a76-75e1b7593199-combined-ca-bundle\") pod \"barbican-worker-fb9947b9-nbdwz\" (UID: \"3958bab3-ce66-4d2d-8a76-75e1b7593199\") " pod="openstack/barbican-worker-fb9947b9-nbdwz" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.939443 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5925d65-97cd-493a-8114-6ed07de4386d-combined-ca-bundle\") pod \"barbican-keystone-listener-5c65499f7d-s6w7z\" (UID: \"a5925d65-97cd-493a-8114-6ed07de4386d\") " pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.940297 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5925d65-97cd-493a-8114-6ed07de4386d-config-data\") pod \"barbican-keystone-listener-5c65499f7d-s6w7z\" (UID: \"a5925d65-97cd-493a-8114-6ed07de4386d\") " pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.942460 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3958bab3-ce66-4d2d-8a76-75e1b7593199-config-data\") pod \"barbican-worker-fb9947b9-nbdwz\" (UID: \"3958bab3-ce66-4d2d-8a76-75e1b7593199\") " pod="openstack/barbican-worker-fb9947b9-nbdwz" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.943498 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mz625\" (UniqueName: \"kubernetes.io/projected/a5925d65-97cd-493a-8114-6ed07de4386d-kube-api-access-mz625\") pod \"barbican-keystone-listener-5c65499f7d-s6w7z\" (UID: \"a5925d65-97cd-493a-8114-6ed07de4386d\") " pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.948027 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-57dfcbc748-z27b5"] Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.950806 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.953086 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.961379 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5925d65-97cd-493a-8114-6ed07de4386d-config-data-custom\") pod \"barbican-keystone-listener-5c65499f7d-s6w7z\" (UID: \"a5925d65-97cd-493a-8114-6ed07de4386d\") " pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.971025 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7k89\" (UniqueName: \"kubernetes.io/projected/3958bab3-ce66-4d2d-8a76-75e1b7593199-kube-api-access-b7k89\") pod \"barbican-worker-fb9947b9-nbdwz\" (UID: \"3958bab3-ce66-4d2d-8a76-75e1b7593199\") " pod="openstack/barbican-worker-fb9947b9-nbdwz" Nov 22 05:11:33 crc kubenswrapper[4818]: I1122 05:11:33.972283 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57dfcbc748-z27b5"] Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.004396 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-config\") pod \"dnsmasq-dns-6bb684768f-rrhr2\" (UID: \"64f9d89e-e211-4ee2-b78f-4321d612a468\") " pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.005562 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-dns-svc\") pod \"dnsmasq-dns-6bb684768f-rrhr2\" (UID: \"64f9d89e-e211-4ee2-b78f-4321d612a468\") " pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.005609 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb684768f-rrhr2\" (UID: \"64f9d89e-e211-4ee2-b78f-4321d612a468\") " pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.005654 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc2feeaa-d88f-4114-9e61-ae0df78e123c-combined-ca-bundle\") pod \"barbican-api-57dfcbc748-z27b5\" (UID: \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\") " pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.005722 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhpkq\" (UniqueName: \"kubernetes.io/projected/fc2feeaa-d88f-4114-9e61-ae0df78e123c-kube-api-access-fhpkq\") pod \"barbican-api-57dfcbc748-z27b5\" (UID: \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\") " pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.005751 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n5j6\" (UniqueName: \"kubernetes.io/projected/64f9d89e-e211-4ee2-b78f-4321d612a468-kube-api-access-8n5j6\") pod \"dnsmasq-dns-6bb684768f-rrhr2\" (UID: \"64f9d89e-e211-4ee2-b78f-4321d612a468\") " pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.005773 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb684768f-rrhr2\" (UID: \"64f9d89e-e211-4ee2-b78f-4321d612a468\") " pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.005853 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc2feeaa-d88f-4114-9e61-ae0df78e123c-config-data-custom\") pod \"barbican-api-57dfcbc748-z27b5\" (UID: \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\") " pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.005875 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc2feeaa-d88f-4114-9e61-ae0df78e123c-config-data\") pod \"barbican-api-57dfcbc748-z27b5\" (UID: \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\") " pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.005902 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc2feeaa-d88f-4114-9e61-ae0df78e123c-logs\") pod \"barbican-api-57dfcbc748-z27b5\" (UID: \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\") " pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.022629 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-fb9947b9-nbdwz" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.045114 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.106938 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-config\") pod \"dnsmasq-dns-6bb684768f-rrhr2\" (UID: \"64f9d89e-e211-4ee2-b78f-4321d612a468\") " pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.107009 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-dns-svc\") pod \"dnsmasq-dns-6bb684768f-rrhr2\" (UID: \"64f9d89e-e211-4ee2-b78f-4321d612a468\") " pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.107044 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb684768f-rrhr2\" (UID: \"64f9d89e-e211-4ee2-b78f-4321d612a468\") " pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.107073 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc2feeaa-d88f-4114-9e61-ae0df78e123c-combined-ca-bundle\") pod \"barbican-api-57dfcbc748-z27b5\" (UID: \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\") " pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.107094 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhpkq\" (UniqueName: \"kubernetes.io/projected/fc2feeaa-d88f-4114-9e61-ae0df78e123c-kube-api-access-fhpkq\") pod \"barbican-api-57dfcbc748-z27b5\" (UID: \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\") " pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.107112 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n5j6\" (UniqueName: \"kubernetes.io/projected/64f9d89e-e211-4ee2-b78f-4321d612a468-kube-api-access-8n5j6\") pod \"dnsmasq-dns-6bb684768f-rrhr2\" (UID: \"64f9d89e-e211-4ee2-b78f-4321d612a468\") " pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.107130 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb684768f-rrhr2\" (UID: \"64f9d89e-e211-4ee2-b78f-4321d612a468\") " pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.107180 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc2feeaa-d88f-4114-9e61-ae0df78e123c-config-data-custom\") pod \"barbican-api-57dfcbc748-z27b5\" (UID: \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\") " pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.107203 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc2feeaa-d88f-4114-9e61-ae0df78e123c-config-data\") pod \"barbican-api-57dfcbc748-z27b5\" (UID: \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\") " pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.107223 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc2feeaa-d88f-4114-9e61-ae0df78e123c-logs\") pod \"barbican-api-57dfcbc748-z27b5\" (UID: \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\") " pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.107879 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-config\") pod \"dnsmasq-dns-6bb684768f-rrhr2\" (UID: \"64f9d89e-e211-4ee2-b78f-4321d612a468\") " pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.108233 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb684768f-rrhr2\" (UID: \"64f9d89e-e211-4ee2-b78f-4321d612a468\") " pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.108280 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-dns-svc\") pod \"dnsmasq-dns-6bb684768f-rrhr2\" (UID: \"64f9d89e-e211-4ee2-b78f-4321d612a468\") " pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.108312 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb684768f-rrhr2\" (UID: \"64f9d89e-e211-4ee2-b78f-4321d612a468\") " pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.108505 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc2feeaa-d88f-4114-9e61-ae0df78e123c-logs\") pod \"barbican-api-57dfcbc748-z27b5\" (UID: \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\") " pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.112040 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc2feeaa-d88f-4114-9e61-ae0df78e123c-combined-ca-bundle\") pod \"barbican-api-57dfcbc748-z27b5\" (UID: \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\") " pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.112697 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc2feeaa-d88f-4114-9e61-ae0df78e123c-config-data\") pod \"barbican-api-57dfcbc748-z27b5\" (UID: \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\") " pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.114130 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc2feeaa-d88f-4114-9e61-ae0df78e123c-config-data-custom\") pod \"barbican-api-57dfcbc748-z27b5\" (UID: \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\") " pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.127667 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhpkq\" (UniqueName: \"kubernetes.io/projected/fc2feeaa-d88f-4114-9e61-ae0df78e123c-kube-api-access-fhpkq\") pod \"barbican-api-57dfcbc748-z27b5\" (UID: \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\") " pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.127964 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n5j6\" (UniqueName: \"kubernetes.io/projected/64f9d89e-e211-4ee2-b78f-4321d612a468-kube-api-access-8n5j6\") pod \"dnsmasq-dns-6bb684768f-rrhr2\" (UID: \"64f9d89e-e211-4ee2-b78f-4321d612a468\") " pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.358378 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.365031 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" podUID="8e82d7f2-2913-48be-ae98-28596edf9d4a" containerName="dnsmasq-dns" containerID="cri-o://1307d498f0f33be235accb53597a8787d249fab7e5cc31b0c95f406288ac414e" gracePeriod=10 Nov 22 05:11:34 crc kubenswrapper[4818]: I1122 05:11:34.367261 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:35 crc kubenswrapper[4818]: I1122 05:11:35.385135 4818 generic.go:334] "Generic (PLEG): container finished" podID="8e82d7f2-2913-48be-ae98-28596edf9d4a" containerID="1307d498f0f33be235accb53597a8787d249fab7e5cc31b0c95f406288ac414e" exitCode=0 Nov 22 05:11:35 crc kubenswrapper[4818]: I1122 05:11:35.385218 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" event={"ID":"8e82d7f2-2913-48be-ae98-28596edf9d4a","Type":"ContainerDied","Data":"1307d498f0f33be235accb53597a8787d249fab7e5cc31b0c95f406288ac414e"} Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.257762 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6dfb48dcf4-6ph2v"] Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.260063 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.262969 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.263038 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.279486 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6dfb48dcf4-6ph2v"] Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.343121 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kktmn\" (UniqueName: \"kubernetes.io/projected/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-kube-api-access-kktmn\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.343196 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-combined-ca-bundle\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.343245 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-logs\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.343301 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-public-tls-certs\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.343348 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-config-data\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.343373 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-internal-tls-certs\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.343451 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-config-data-custom\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.446076 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-logs\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.446142 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-public-tls-certs\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.446200 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-config-data\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.446233 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-internal-tls-certs\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.446373 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-config-data-custom\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.446654 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kktmn\" (UniqueName: \"kubernetes.io/projected/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-kube-api-access-kktmn\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.446714 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-combined-ca-bundle\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.447344 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-logs\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.517201 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-combined-ca-bundle\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.550056 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-internal-tls-certs\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.551025 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-config-data-custom\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.552023 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-public-tls-certs\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.554294 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-config-data\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.560394 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kktmn\" (UniqueName: \"kubernetes.io/projected/1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d-kube-api-access-kktmn\") pod \"barbican-api-6dfb48dcf4-6ph2v\" (UID: \"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d\") " pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.625776 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.724976 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.829728 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-fb9947b9-nbdwz"] Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.857708 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-ovsdbserver-nb\") pod \"8e82d7f2-2913-48be-ae98-28596edf9d4a\" (UID: \"8e82d7f2-2913-48be-ae98-28596edf9d4a\") " Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.857772 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-dns-svc\") pod \"8e82d7f2-2913-48be-ae98-28596edf9d4a\" (UID: \"8e82d7f2-2913-48be-ae98-28596edf9d4a\") " Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.857840 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9s77\" (UniqueName: \"kubernetes.io/projected/8e82d7f2-2913-48be-ae98-28596edf9d4a-kube-api-access-q9s77\") pod \"8e82d7f2-2913-48be-ae98-28596edf9d4a\" (UID: \"8e82d7f2-2913-48be-ae98-28596edf9d4a\") " Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.857933 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-config\") pod \"8e82d7f2-2913-48be-ae98-28596edf9d4a\" (UID: \"8e82d7f2-2913-48be-ae98-28596edf9d4a\") " Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.857988 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-ovsdbserver-sb\") pod \"8e82d7f2-2913-48be-ae98-28596edf9d4a\" (UID: \"8e82d7f2-2913-48be-ae98-28596edf9d4a\") " Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.873467 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e82d7f2-2913-48be-ae98-28596edf9d4a-kube-api-access-q9s77" (OuterVolumeSpecName: "kube-api-access-q9s77") pod "8e82d7f2-2913-48be-ae98-28596edf9d4a" (UID: "8e82d7f2-2913-48be-ae98-28596edf9d4a"). InnerVolumeSpecName "kube-api-access-q9s77". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.905375 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8e82d7f2-2913-48be-ae98-28596edf9d4a" (UID: "8e82d7f2-2913-48be-ae98-28596edf9d4a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.914543 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8e82d7f2-2913-48be-ae98-28596edf9d4a" (UID: "8e82d7f2-2913-48be-ae98-28596edf9d4a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.922184 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8e82d7f2-2913-48be-ae98-28596edf9d4a" (UID: "8e82d7f2-2913-48be-ae98-28596edf9d4a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.924338 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-config" (OuterVolumeSpecName: "config") pod "8e82d7f2-2913-48be-ae98-28596edf9d4a" (UID: "8e82d7f2-2913-48be-ae98-28596edf9d4a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.961348 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.961387 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.961400 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9s77\" (UniqueName: \"kubernetes.io/projected/8e82d7f2-2913-48be-ae98-28596edf9d4a-kube-api-access-q9s77\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.961413 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:36 crc kubenswrapper[4818]: I1122 05:11:36.961422 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e82d7f2-2913-48be-ae98-28596edf9d4a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:37 crc kubenswrapper[4818]: I1122 05:11:37.010531 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5c65499f7d-s6w7z"] Nov 22 05:11:37 crc kubenswrapper[4818]: I1122 05:11:37.022704 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-rrhr2"] Nov 22 05:11:37 crc kubenswrapper[4818]: I1122 05:11:37.201646 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57dfcbc748-z27b5"] Nov 22 05:11:37 crc kubenswrapper[4818]: I1122 05:11:37.216170 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6dfb48dcf4-6ph2v"] Nov 22 05:11:37 crc kubenswrapper[4818]: W1122 05:11:37.240693 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc2feeaa_d88f_4114_9e61_ae0df78e123c.slice/crio-5d23ff50d49decdb3391c830a39d665163994acf9167d7c3da9b148f233b2c5e WatchSource:0}: Error finding container 5d23ff50d49decdb3391c830a39d665163994acf9167d7c3da9b148f233b2c5e: Status 404 returned error can't find the container with id 5d23ff50d49decdb3391c830a39d665163994acf9167d7c3da9b148f233b2c5e Nov 22 05:11:37 crc kubenswrapper[4818]: W1122 05:11:37.261870 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e1ed319_7d7b_4a93_925c_c7bf4bbfaa1d.slice/crio-cd791d195139d81a33b23e852337d820089b449fe82eb3a72049ad74b925e2bb WatchSource:0}: Error finding container cd791d195139d81a33b23e852337d820089b449fe82eb3a72049ad74b925e2bb: Status 404 returned error can't find the container with id cd791d195139d81a33b23e852337d820089b449fe82eb3a72049ad74b925e2bb Nov 22 05:11:37 crc kubenswrapper[4818]: I1122 05:11:37.407573 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" event={"ID":"a5925d65-97cd-493a-8114-6ed07de4386d","Type":"ContainerStarted","Data":"aa3b8bfe5d6418aa6ae66f7763b2a0ba84b1b43c38e72518ab8d4fd17b8efb4a"} Nov 22 05:11:37 crc kubenswrapper[4818]: I1122 05:11:37.409116 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-fb9947b9-nbdwz" event={"ID":"3958bab3-ce66-4d2d-8a76-75e1b7593199","Type":"ContainerStarted","Data":"dc36a77eb6922e6f76de8701bb0ed1dc04bde79423c53d677258c7764172afcb"} Nov 22 05:11:37 crc kubenswrapper[4818]: I1122 05:11:37.410553 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6dfb48dcf4-6ph2v" event={"ID":"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d","Type":"ContainerStarted","Data":"cd791d195139d81a33b23e852337d820089b449fe82eb3a72049ad74b925e2bb"} Nov 22 05:11:37 crc kubenswrapper[4818]: I1122 05:11:37.412025 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57dfcbc748-z27b5" event={"ID":"fc2feeaa-d88f-4114-9e61-ae0df78e123c","Type":"ContainerStarted","Data":"5d23ff50d49decdb3391c830a39d665163994acf9167d7c3da9b148f233b2c5e"} Nov 22 05:11:37 crc kubenswrapper[4818]: I1122 05:11:37.417912 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" event={"ID":"8e82d7f2-2913-48be-ae98-28596edf9d4a","Type":"ContainerDied","Data":"27072eaa1ea39a5268e2f28a3ca4257df6fb79d3a809abded046fab16f024099"} Nov 22 05:11:37 crc kubenswrapper[4818]: I1122 05:11:37.417961 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b946d459c-ts7sf" Nov 22 05:11:37 crc kubenswrapper[4818]: I1122 05:11:37.417979 4818 scope.go:117] "RemoveContainer" containerID="1307d498f0f33be235accb53597a8787d249fab7e5cc31b0c95f406288ac414e" Nov 22 05:11:37 crc kubenswrapper[4818]: I1122 05:11:37.420267 4818 generic.go:334] "Generic (PLEG): container finished" podID="64f9d89e-e211-4ee2-b78f-4321d612a468" containerID="4b86475529269b6b0fe1f5baa49a5f8ffc521586f2ecf14ad57ebdc5de7abe36" exitCode=0 Nov 22 05:11:37 crc kubenswrapper[4818]: I1122 05:11:37.420367 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" event={"ID":"64f9d89e-e211-4ee2-b78f-4321d612a468","Type":"ContainerDied","Data":"4b86475529269b6b0fe1f5baa49a5f8ffc521586f2ecf14ad57ebdc5de7abe36"} Nov 22 05:11:37 crc kubenswrapper[4818]: I1122 05:11:37.420503 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" event={"ID":"64f9d89e-e211-4ee2-b78f-4321d612a468","Type":"ContainerStarted","Data":"9c0b77b889e2f91021ff43dddb9574ce01f54a3ae0b354701c3b70b94c0ee3ce"} Nov 22 05:11:37 crc kubenswrapper[4818]: I1122 05:11:37.422218 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c105a582-4a5b-4c1b-aac3-bbde151fcb1b","Type":"ContainerStarted","Data":"b522dea24eec23301001232d2f761e1aca84d0a3f33389e65a611df5534e9e3a"} Nov 22 05:11:37 crc kubenswrapper[4818]: I1122 05:11:37.453668 4818 scope.go:117] "RemoveContainer" containerID="fa16ed87ba70a91bb87f1391b24f651455dc50a14bf20b9b1de2ca2f53026cb9" Nov 22 05:11:37 crc kubenswrapper[4818]: I1122 05:11:37.543711 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-ts7sf"] Nov 22 05:11:37 crc kubenswrapper[4818]: I1122 05:11:37.549646 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-ts7sf"] Nov 22 05:11:38 crc kubenswrapper[4818]: I1122 05:11:38.333074 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e82d7f2-2913-48be-ae98-28596edf9d4a" path="/var/lib/kubelet/pods/8e82d7f2-2913-48be-ae98-28596edf9d4a/volumes" Nov 22 05:11:38 crc kubenswrapper[4818]: I1122 05:11:38.443668 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c105a582-4a5b-4c1b-aac3-bbde151fcb1b","Type":"ContainerStarted","Data":"4f4db4c8ecf36d0202f1a721660dec3efe45b3d18072d5cd524421af404e3daa"} Nov 22 05:11:38 crc kubenswrapper[4818]: I1122 05:11:38.446826 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" event={"ID":"64f9d89e-e211-4ee2-b78f-4321d612a468","Type":"ContainerStarted","Data":"6a0d11f3a54a4cb8b10b17618fd3ee114699b87b7f44a5e445ab8ffa8bdfe6d5"} Nov 22 05:11:38 crc kubenswrapper[4818]: I1122 05:11:38.447057 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:38 crc kubenswrapper[4818]: I1122 05:11:38.468390 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6dfb48dcf4-6ph2v" event={"ID":"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d","Type":"ContainerStarted","Data":"11488518d5648bb168425728c9623524ffe60a6c25656bfbea80140198d19617"} Nov 22 05:11:38 crc kubenswrapper[4818]: I1122 05:11:38.468439 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6dfb48dcf4-6ph2v" event={"ID":"1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d","Type":"ContainerStarted","Data":"f1bd0d9e333cc82a460ffcf77be4a754a4dab70581151b14717cfe360548572a"} Nov 22 05:11:38 crc kubenswrapper[4818]: I1122 05:11:38.469201 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:38 crc kubenswrapper[4818]: I1122 05:11:38.469226 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:38 crc kubenswrapper[4818]: I1122 05:11:38.476092 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" podStartSLOduration=5.476069162 podStartE2EDuration="5.476069162s" podCreationTimestamp="2025-11-22 05:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:11:38.469533519 +0000 UTC m=+1451.043950046" watchObservedRunningTime="2025-11-22 05:11:38.476069162 +0000 UTC m=+1451.050485689" Nov 22 05:11:38 crc kubenswrapper[4818]: I1122 05:11:38.476325 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57dfcbc748-z27b5" event={"ID":"fc2feeaa-d88f-4114-9e61-ae0df78e123c","Type":"ContainerStarted","Data":"61982d5f0de0f9899cc92f39d18fb52198b8bf6b87fdba3371b4d2182c507c44"} Nov 22 05:11:38 crc kubenswrapper[4818]: I1122 05:11:38.476360 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57dfcbc748-z27b5" event={"ID":"fc2feeaa-d88f-4114-9e61-ae0df78e123c","Type":"ContainerStarted","Data":"bf53fb270cfa817e4ab79364a9d5d6b7a4303fa2679fdb75110f922d09189d08"} Nov 22 05:11:38 crc kubenswrapper[4818]: I1122 05:11:38.476647 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:38 crc kubenswrapper[4818]: I1122 05:11:38.477034 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:38 crc kubenswrapper[4818]: I1122 05:11:38.497544 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6dfb48dcf4-6ph2v" podStartSLOduration=2.497525888 podStartE2EDuration="2.497525888s" podCreationTimestamp="2025-11-22 05:11:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:11:38.49227975 +0000 UTC m=+1451.066696297" watchObservedRunningTime="2025-11-22 05:11:38.497525888 +0000 UTC m=+1451.071942415" Nov 22 05:11:38 crc kubenswrapper[4818]: I1122 05:11:38.518841 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-57dfcbc748-z27b5" podStartSLOduration=5.518816 podStartE2EDuration="5.518816s" podCreationTimestamp="2025-11-22 05:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:11:38.512075142 +0000 UTC m=+1451.086491669" watchObservedRunningTime="2025-11-22 05:11:38.518816 +0000 UTC m=+1451.093232527" Nov 22 05:11:40 crc kubenswrapper[4818]: I1122 05:11:40.499489 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" event={"ID":"a5925d65-97cd-493a-8114-6ed07de4386d","Type":"ContainerStarted","Data":"aa9a8e8564b6305097d38a0040942cf22151795b6064406180057289a4c269a7"} Nov 22 05:11:40 crc kubenswrapper[4818]: I1122 05:11:40.500126 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" event={"ID":"a5925d65-97cd-493a-8114-6ed07de4386d","Type":"ContainerStarted","Data":"096d10df4c7194ac2aa4ad47f7f42d9a9d457a8c59fd942b3412e54b5261137e"} Nov 22 05:11:40 crc kubenswrapper[4818]: I1122 05:11:40.504575 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-fb9947b9-nbdwz" event={"ID":"3958bab3-ce66-4d2d-8a76-75e1b7593199","Type":"ContainerStarted","Data":"5f35d1d74724aeccac4a46006ae8eb2d2e8a6aa1e62351888f0f3c020ee40581"} Nov 22 05:11:40 crc kubenswrapper[4818]: I1122 05:11:40.505431 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-fb9947b9-nbdwz" event={"ID":"3958bab3-ce66-4d2d-8a76-75e1b7593199","Type":"ContainerStarted","Data":"9cb5532b273acd905fed3ae1faee25491cd38203d83fa03e79c41068a61e38a8"} Nov 22 05:11:40 crc kubenswrapper[4818]: I1122 05:11:40.511802 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c105a582-4a5b-4c1b-aac3-bbde151fcb1b","Type":"ContainerStarted","Data":"440f38a06792aeef09e5e5c5d8c5e80dee11e1a69f69f96e2daf060a4807dbf0"} Nov 22 05:11:40 crc kubenswrapper[4818]: I1122 05:11:40.528579 4818 generic.go:334] "Generic (PLEG): container finished" podID="0ca1db80-7d8d-44cf-80bf-afa1c0798182" containerID="fed7bf830da4016586061964fd8e2a8b04bd20aa622442114222ec6e1297420a" exitCode=0 Nov 22 05:11:40 crc kubenswrapper[4818]: I1122 05:11:40.528910 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7spb7" event={"ID":"0ca1db80-7d8d-44cf-80bf-afa1c0798182","Type":"ContainerDied","Data":"fed7bf830da4016586061964fd8e2a8b04bd20aa622442114222ec6e1297420a"} Nov 22 05:11:40 crc kubenswrapper[4818]: I1122 05:11:40.533610 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5c65499f7d-s6w7z" podStartSLOduration=5.202493668 podStartE2EDuration="7.533590795s" podCreationTimestamp="2025-11-22 05:11:33 +0000 UTC" firstStartedPulling="2025-11-22 05:11:37.010913989 +0000 UTC m=+1449.585330516" lastFinishedPulling="2025-11-22 05:11:39.342011096 +0000 UTC m=+1451.916427643" observedRunningTime="2025-11-22 05:11:40.528556802 +0000 UTC m=+1453.102973339" watchObservedRunningTime="2025-11-22 05:11:40.533590795 +0000 UTC m=+1453.108007322" Nov 22 05:11:40 crc kubenswrapper[4818]: I1122 05:11:40.566202 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-fb9947b9-nbdwz" podStartSLOduration=5.08690089 podStartE2EDuration="7.566179584s" podCreationTimestamp="2025-11-22 05:11:33 +0000 UTC" firstStartedPulling="2025-11-22 05:11:36.833424089 +0000 UTC m=+1449.407840616" lastFinishedPulling="2025-11-22 05:11:39.312702773 +0000 UTC m=+1451.887119310" observedRunningTime="2025-11-22 05:11:40.55846892 +0000 UTC m=+1453.132885467" watchObservedRunningTime="2025-11-22 05:11:40.566179584 +0000 UTC m=+1453.140596121" Nov 22 05:11:41 crc kubenswrapper[4818]: I1122 05:11:41.602749 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c105a582-4a5b-4c1b-aac3-bbde151fcb1b","Type":"ContainerStarted","Data":"5a9ef5cee36644f0cb6cd57d370f33f472dc1aaed14c909eda40f0a18dc88eea"} Nov 22 05:11:41 crc kubenswrapper[4818]: I1122 05:11:41.603612 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 05:11:41 crc kubenswrapper[4818]: I1122 05:11:41.646548 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.145987969 podStartE2EDuration="10.646529511s" podCreationTimestamp="2025-11-22 05:11:31 +0000 UTC" firstStartedPulling="2025-11-22 05:11:32.330182128 +0000 UTC m=+1444.904598675" lastFinishedPulling="2025-11-22 05:11:40.83072369 +0000 UTC m=+1453.405140217" observedRunningTime="2025-11-22 05:11:41.642755521 +0000 UTC m=+1454.217172048" watchObservedRunningTime="2025-11-22 05:11:41.646529511 +0000 UTC m=+1454.220946038" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.062055 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7spb7" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.174899 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-db-sync-config-data\") pod \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.175061 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-combined-ca-bundle\") pod \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.175095 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-scripts\") pod \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.175156 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0ca1db80-7d8d-44cf-80bf-afa1c0798182-etc-machine-id\") pod \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.175189 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-config-data\") pod \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.175217 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwvrh\" (UniqueName: \"kubernetes.io/projected/0ca1db80-7d8d-44cf-80bf-afa1c0798182-kube-api-access-nwvrh\") pod \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\" (UID: \"0ca1db80-7d8d-44cf-80bf-afa1c0798182\") " Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.178029 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0ca1db80-7d8d-44cf-80bf-afa1c0798182-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0ca1db80-7d8d-44cf-80bf-afa1c0798182" (UID: "0ca1db80-7d8d-44cf-80bf-afa1c0798182"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.181054 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-scripts" (OuterVolumeSpecName: "scripts") pod "0ca1db80-7d8d-44cf-80bf-afa1c0798182" (UID: "0ca1db80-7d8d-44cf-80bf-afa1c0798182"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.181626 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0ca1db80-7d8d-44cf-80bf-afa1c0798182" (UID: "0ca1db80-7d8d-44cf-80bf-afa1c0798182"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.181795 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ca1db80-7d8d-44cf-80bf-afa1c0798182-kube-api-access-nwvrh" (OuterVolumeSpecName: "kube-api-access-nwvrh") pod "0ca1db80-7d8d-44cf-80bf-afa1c0798182" (UID: "0ca1db80-7d8d-44cf-80bf-afa1c0798182"). InnerVolumeSpecName "kube-api-access-nwvrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.205003 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ca1db80-7d8d-44cf-80bf-afa1c0798182" (UID: "0ca1db80-7d8d-44cf-80bf-afa1c0798182"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.258825 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-config-data" (OuterVolumeSpecName: "config-data") pod "0ca1db80-7d8d-44cf-80bf-afa1c0798182" (UID: "0ca1db80-7d8d-44cf-80bf-afa1c0798182"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.277106 4818 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.277155 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.277173 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.277187 4818 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0ca1db80-7d8d-44cf-80bf-afa1c0798182-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.277201 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ca1db80-7d8d-44cf-80bf-afa1c0798182-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.277215 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwvrh\" (UniqueName: \"kubernetes.io/projected/0ca1db80-7d8d-44cf-80bf-afa1c0798182-kube-api-access-nwvrh\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.629225 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7spb7" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.630128 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7spb7" event={"ID":"0ca1db80-7d8d-44cf-80bf-afa1c0798182","Type":"ContainerDied","Data":"a017ef0933330bf734fddc35ecf89b4f6e30c951af50a73dd41969f663162b68"} Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.630171 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a017ef0933330bf734fddc35ecf89b4f6e30c951af50a73dd41969f663162b68" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.900183 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 05:11:42 crc kubenswrapper[4818]: E1122 05:11:42.900538 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e82d7f2-2913-48be-ae98-28596edf9d4a" containerName="dnsmasq-dns" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.900554 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e82d7f2-2913-48be-ae98-28596edf9d4a" containerName="dnsmasq-dns" Nov 22 05:11:42 crc kubenswrapper[4818]: E1122 05:11:42.900578 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca1db80-7d8d-44cf-80bf-afa1c0798182" containerName="cinder-db-sync" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.900586 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca1db80-7d8d-44cf-80bf-afa1c0798182" containerName="cinder-db-sync" Nov 22 05:11:42 crc kubenswrapper[4818]: E1122 05:11:42.900603 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e82d7f2-2913-48be-ae98-28596edf9d4a" containerName="init" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.900609 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e82d7f2-2913-48be-ae98-28596edf9d4a" containerName="init" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.900775 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ca1db80-7d8d-44cf-80bf-afa1c0798182" containerName="cinder-db-sync" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.900793 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e82d7f2-2913-48be-ae98-28596edf9d4a" containerName="dnsmasq-dns" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.901669 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.907593 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.907680 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-9pzqx" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.907918 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.908138 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.933322 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.971400 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-rrhr2"] Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.971669 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" podUID="64f9d89e-e211-4ee2-b78f-4321d612a468" containerName="dnsmasq-dns" containerID="cri-o://6a0d11f3a54a4cb8b10b17618fd3ee114699b87b7f44a5e445ab8ffa8bdfe6d5" gracePeriod=10 Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.972743 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.990165 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-scripts\") pod \"cinder-scheduler-0\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " pod="openstack/cinder-scheduler-0" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.990204 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-config-data\") pod \"cinder-scheduler-0\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " pod="openstack/cinder-scheduler-0" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.990222 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " pod="openstack/cinder-scheduler-0" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.990308 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c695f78-88f9-4013-a67b-4148203ff826-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " pod="openstack/cinder-scheduler-0" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.990336 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " pod="openstack/cinder-scheduler-0" Nov 22 05:11:42 crc kubenswrapper[4818]: I1122 05:11:42.990364 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5pnt\" (UniqueName: \"kubernetes.io/projected/6c695f78-88f9-4013-a67b-4148203ff826-kube-api-access-h5pnt\") pod \"cinder-scheduler-0\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " pod="openstack/cinder-scheduler-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.037030 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-dh7x5"] Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.038621 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.051245 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-dh7x5"] Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.091856 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-scripts\") pod \"cinder-scheduler-0\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " pod="openstack/cinder-scheduler-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.092998 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-config-data\") pod \"cinder-scheduler-0\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " pod="openstack/cinder-scheduler-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.093037 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " pod="openstack/cinder-scheduler-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.093093 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c695f78-88f9-4013-a67b-4148203ff826-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " pod="openstack/cinder-scheduler-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.093147 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " pod="openstack/cinder-scheduler-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.093203 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5pnt\" (UniqueName: \"kubernetes.io/projected/6c695f78-88f9-4013-a67b-4148203ff826-kube-api-access-h5pnt\") pod \"cinder-scheduler-0\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " pod="openstack/cinder-scheduler-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.093912 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c695f78-88f9-4013-a67b-4148203ff826-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " pod="openstack/cinder-scheduler-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.097268 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-scripts\") pod \"cinder-scheduler-0\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " pod="openstack/cinder-scheduler-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.099477 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-config-data\") pod \"cinder-scheduler-0\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " pod="openstack/cinder-scheduler-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.101865 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " pod="openstack/cinder-scheduler-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.102942 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " pod="openstack/cinder-scheduler-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.123100 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5pnt\" (UniqueName: \"kubernetes.io/projected/6c695f78-88f9-4013-a67b-4148203ff826-kube-api-access-h5pnt\") pod \"cinder-scheduler-0\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " pod="openstack/cinder-scheduler-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.181218 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.182720 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.186601 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.200481 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-config\") pod \"dnsmasq-dns-6d97fcdd8f-dh7x5\" (UID: \"a527f913-2977-419c-9a4c-25b4c1729025\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.200536 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-ovsdbserver-sb\") pod \"dnsmasq-dns-6d97fcdd8f-dh7x5\" (UID: \"a527f913-2977-419c-9a4c-25b4c1729025\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.200568 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-dns-svc\") pod \"dnsmasq-dns-6d97fcdd8f-dh7x5\" (UID: \"a527f913-2977-419c-9a4c-25b4c1729025\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.200588 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-ovsdbserver-nb\") pod \"dnsmasq-dns-6d97fcdd8f-dh7x5\" (UID: \"a527f913-2977-419c-9a4c-25b4c1729025\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.200625 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v76vm\" (UniqueName: \"kubernetes.io/projected/a527f913-2977-419c-9a4c-25b4c1729025-kube-api-access-v76vm\") pod \"dnsmasq-dns-6d97fcdd8f-dh7x5\" (UID: \"a527f913-2977-419c-9a4c-25b4c1729025\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.201153 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.223346 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.302648 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d760b6-0b81-4e56-855b-c54f8db80db7-logs\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.302762 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.302820 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-config-data-custom\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.302916 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-config\") pod \"dnsmasq-dns-6d97fcdd8f-dh7x5\" (UID: \"a527f913-2977-419c-9a4c-25b4c1729025\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.302944 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-scripts\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.302990 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jd49\" (UniqueName: \"kubernetes.io/projected/23d760b6-0b81-4e56-855b-c54f8db80db7-kube-api-access-4jd49\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.303046 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-ovsdbserver-sb\") pod \"dnsmasq-dns-6d97fcdd8f-dh7x5\" (UID: \"a527f913-2977-419c-9a4c-25b4c1729025\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.303102 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-dns-svc\") pod \"dnsmasq-dns-6d97fcdd8f-dh7x5\" (UID: \"a527f913-2977-419c-9a4c-25b4c1729025\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.303138 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-ovsdbserver-nb\") pod \"dnsmasq-dns-6d97fcdd8f-dh7x5\" (UID: \"a527f913-2977-419c-9a4c-25b4c1729025\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.303200 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23d760b6-0b81-4e56-855b-c54f8db80db7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.303268 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v76vm\" (UniqueName: \"kubernetes.io/projected/a527f913-2977-419c-9a4c-25b4c1729025-kube-api-access-v76vm\") pod \"dnsmasq-dns-6d97fcdd8f-dh7x5\" (UID: \"a527f913-2977-419c-9a4c-25b4c1729025\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.303360 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-config-data\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.304211 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-config\") pod \"dnsmasq-dns-6d97fcdd8f-dh7x5\" (UID: \"a527f913-2977-419c-9a4c-25b4c1729025\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.307794 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-ovsdbserver-nb\") pod \"dnsmasq-dns-6d97fcdd8f-dh7x5\" (UID: \"a527f913-2977-419c-9a4c-25b4c1729025\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.314698 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-dns-svc\") pod \"dnsmasq-dns-6d97fcdd8f-dh7x5\" (UID: \"a527f913-2977-419c-9a4c-25b4c1729025\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.315817 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-ovsdbserver-sb\") pod \"dnsmasq-dns-6d97fcdd8f-dh7x5\" (UID: \"a527f913-2977-419c-9a4c-25b4c1729025\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.322864 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v76vm\" (UniqueName: \"kubernetes.io/projected/a527f913-2977-419c-9a4c-25b4c1729025-kube-api-access-v76vm\") pod \"dnsmasq-dns-6d97fcdd8f-dh7x5\" (UID: \"a527f913-2977-419c-9a4c-25b4c1729025\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.362529 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.405067 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23d760b6-0b81-4e56-855b-c54f8db80db7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.405450 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-config-data\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.405535 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d760b6-0b81-4e56-855b-c54f8db80db7-logs\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.405562 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.405585 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-config-data-custom\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.405615 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-scripts\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.405642 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jd49\" (UniqueName: \"kubernetes.io/projected/23d760b6-0b81-4e56-855b-c54f8db80db7-kube-api-access-4jd49\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.405220 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23d760b6-0b81-4e56-855b-c54f8db80db7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.409709 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d760b6-0b81-4e56-855b-c54f8db80db7-logs\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.440159 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-config-data-custom\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.440612 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-config-data\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.448839 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jd49\" (UniqueName: \"kubernetes.io/projected/23d760b6-0b81-4e56-855b-c54f8db80db7-kube-api-access-4jd49\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.448876 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-scripts\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.448941 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.514875 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.651815 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.672976 4818 generic.go:334] "Generic (PLEG): container finished" podID="64f9d89e-e211-4ee2-b78f-4321d612a468" containerID="6a0d11f3a54a4cb8b10b17618fd3ee114699b87b7f44a5e445ab8ffa8bdfe6d5" exitCode=0 Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.674064 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.674248 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" event={"ID":"64f9d89e-e211-4ee2-b78f-4321d612a468","Type":"ContainerDied","Data":"6a0d11f3a54a4cb8b10b17618fd3ee114699b87b7f44a5e445ab8ffa8bdfe6d5"} Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.674305 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb684768f-rrhr2" event={"ID":"64f9d89e-e211-4ee2-b78f-4321d612a468","Type":"ContainerDied","Data":"9c0b77b889e2f91021ff43dddb9574ce01f54a3ae0b354701c3b70b94c0ee3ce"} Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.674322 4818 scope.go:117] "RemoveContainer" containerID="6a0d11f3a54a4cb8b10b17618fd3ee114699b87b7f44a5e445ab8ffa8bdfe6d5" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.717050 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-ovsdbserver-sb\") pod \"64f9d89e-e211-4ee2-b78f-4321d612a468\" (UID: \"64f9d89e-e211-4ee2-b78f-4321d612a468\") " Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.717171 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8n5j6\" (UniqueName: \"kubernetes.io/projected/64f9d89e-e211-4ee2-b78f-4321d612a468-kube-api-access-8n5j6\") pod \"64f9d89e-e211-4ee2-b78f-4321d612a468\" (UID: \"64f9d89e-e211-4ee2-b78f-4321d612a468\") " Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.717217 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-dns-svc\") pod \"64f9d89e-e211-4ee2-b78f-4321d612a468\" (UID: \"64f9d89e-e211-4ee2-b78f-4321d612a468\") " Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.717314 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-ovsdbserver-nb\") pod \"64f9d89e-e211-4ee2-b78f-4321d612a468\" (UID: \"64f9d89e-e211-4ee2-b78f-4321d612a468\") " Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.717349 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-config\") pod \"64f9d89e-e211-4ee2-b78f-4321d612a468\" (UID: \"64f9d89e-e211-4ee2-b78f-4321d612a468\") " Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.728669 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64f9d89e-e211-4ee2-b78f-4321d612a468-kube-api-access-8n5j6" (OuterVolumeSpecName: "kube-api-access-8n5j6") pod "64f9d89e-e211-4ee2-b78f-4321d612a468" (UID: "64f9d89e-e211-4ee2-b78f-4321d612a468"). InnerVolumeSpecName "kube-api-access-8n5j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.775662 4818 scope.go:117] "RemoveContainer" containerID="4b86475529269b6b0fe1f5baa49a5f8ffc521586f2ecf14ad57ebdc5de7abe36" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.781871 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-config" (OuterVolumeSpecName: "config") pod "64f9d89e-e211-4ee2-b78f-4321d612a468" (UID: "64f9d89e-e211-4ee2-b78f-4321d612a468"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.784527 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "64f9d89e-e211-4ee2-b78f-4321d612a468" (UID: "64f9d89e-e211-4ee2-b78f-4321d612a468"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.803828 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "64f9d89e-e211-4ee2-b78f-4321d612a468" (UID: "64f9d89e-e211-4ee2-b78f-4321d612a468"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.815772 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.819766 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.819792 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8n5j6\" (UniqueName: \"kubernetes.io/projected/64f9d89e-e211-4ee2-b78f-4321d612a468-kube-api-access-8n5j6\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.819803 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.819811 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.821561 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "64f9d89e-e211-4ee2-b78f-4321d612a468" (UID: "64f9d89e-e211-4ee2-b78f-4321d612a468"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.847541 4818 scope.go:117] "RemoveContainer" containerID="6a0d11f3a54a4cb8b10b17618fd3ee114699b87b7f44a5e445ab8ffa8bdfe6d5" Nov 22 05:11:43 crc kubenswrapper[4818]: E1122 05:11:43.848334 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a0d11f3a54a4cb8b10b17618fd3ee114699b87b7f44a5e445ab8ffa8bdfe6d5\": container with ID starting with 6a0d11f3a54a4cb8b10b17618fd3ee114699b87b7f44a5e445ab8ffa8bdfe6d5 not found: ID does not exist" containerID="6a0d11f3a54a4cb8b10b17618fd3ee114699b87b7f44a5e445ab8ffa8bdfe6d5" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.848528 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a0d11f3a54a4cb8b10b17618fd3ee114699b87b7f44a5e445ab8ffa8bdfe6d5"} err="failed to get container status \"6a0d11f3a54a4cb8b10b17618fd3ee114699b87b7f44a5e445ab8ffa8bdfe6d5\": rpc error: code = NotFound desc = could not find container \"6a0d11f3a54a4cb8b10b17618fd3ee114699b87b7f44a5e445ab8ffa8bdfe6d5\": container with ID starting with 6a0d11f3a54a4cb8b10b17618fd3ee114699b87b7f44a5e445ab8ffa8bdfe6d5 not found: ID does not exist" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.848622 4818 scope.go:117] "RemoveContainer" containerID="4b86475529269b6b0fe1f5baa49a5f8ffc521586f2ecf14ad57ebdc5de7abe36" Nov 22 05:11:43 crc kubenswrapper[4818]: E1122 05:11:43.849016 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b86475529269b6b0fe1f5baa49a5f8ffc521586f2ecf14ad57ebdc5de7abe36\": container with ID starting with 4b86475529269b6b0fe1f5baa49a5f8ffc521586f2ecf14ad57ebdc5de7abe36 not found: ID does not exist" containerID="4b86475529269b6b0fe1f5baa49a5f8ffc521586f2ecf14ad57ebdc5de7abe36" Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.849103 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b86475529269b6b0fe1f5baa49a5f8ffc521586f2ecf14ad57ebdc5de7abe36"} err="failed to get container status \"4b86475529269b6b0fe1f5baa49a5f8ffc521586f2ecf14ad57ebdc5de7abe36\": rpc error: code = NotFound desc = could not find container \"4b86475529269b6b0fe1f5baa49a5f8ffc521586f2ecf14ad57ebdc5de7abe36\": container with ID starting with 4b86475529269b6b0fe1f5baa49a5f8ffc521586f2ecf14ad57ebdc5de7abe36 not found: ID does not exist" Nov 22 05:11:43 crc kubenswrapper[4818]: W1122 05:11:43.865914 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c695f78_88f9_4013_a67b_4148203ff826.slice/crio-4359e0f2ce46891b93b796ed533d131540d570be39d27a3872e2b9000ce14ef5 WatchSource:0}: Error finding container 4359e0f2ce46891b93b796ed533d131540d570be39d27a3872e2b9000ce14ef5: Status 404 returned error can't find the container with id 4359e0f2ce46891b93b796ed533d131540d570be39d27a3872e2b9000ce14ef5 Nov 22 05:11:43 crc kubenswrapper[4818]: I1122 05:11:43.921155 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64f9d89e-e211-4ee2-b78f-4321d612a468-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:44 crc kubenswrapper[4818]: I1122 05:11:44.027846 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-rrhr2"] Nov 22 05:11:44 crc kubenswrapper[4818]: I1122 05:11:44.040380 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-rrhr2"] Nov 22 05:11:44 crc kubenswrapper[4818]: I1122 05:11:44.128825 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-dh7x5"] Nov 22 05:11:44 crc kubenswrapper[4818]: I1122 05:11:44.179016 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 22 05:11:44 crc kubenswrapper[4818]: I1122 05:11:44.317928 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64f9d89e-e211-4ee2-b78f-4321d612a468" path="/var/lib/kubelet/pods/64f9d89e-e211-4ee2-b78f-4321d612a468/volumes" Nov 22 05:11:44 crc kubenswrapper[4818]: I1122 05:11:44.701712 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6c695f78-88f9-4013-a67b-4148203ff826","Type":"ContainerStarted","Data":"4359e0f2ce46891b93b796ed533d131540d570be39d27a3872e2b9000ce14ef5"} Nov 22 05:11:44 crc kubenswrapper[4818]: I1122 05:11:44.704113 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"23d760b6-0b81-4e56-855b-c54f8db80db7","Type":"ContainerStarted","Data":"82e6291f33c21e6384845395ceed938e4b26b04ef9a0b053a6c776d419d81ece"} Nov 22 05:11:44 crc kubenswrapper[4818]: I1122 05:11:44.711321 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" event={"ID":"a527f913-2977-419c-9a4c-25b4c1729025","Type":"ContainerStarted","Data":"a85c8dbbfe43d1f5977724516652d8cd47242546de587eac7870ab7585c6c0f7"} Nov 22 05:11:44 crc kubenswrapper[4818]: I1122 05:11:44.711355 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" event={"ID":"a527f913-2977-419c-9a4c-25b4c1729025","Type":"ContainerStarted","Data":"7c370fad63c7f518851650c725bc3157cdcbee262d848a6db0078aa6dd1b56be"} Nov 22 05:11:45 crc kubenswrapper[4818]: I1122 05:11:45.254458 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 22 05:11:45 crc kubenswrapper[4818]: I1122 05:11:45.723631 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"23d760b6-0b81-4e56-855b-c54f8db80db7","Type":"ContainerStarted","Data":"8a93f7f1f40f37da2a06bca4fd23a35c988571ff1dbe202979fbc17c43d3116a"} Nov 22 05:11:45 crc kubenswrapper[4818]: I1122 05:11:45.726491 4818 generic.go:334] "Generic (PLEG): container finished" podID="a527f913-2977-419c-9a4c-25b4c1729025" containerID="a85c8dbbfe43d1f5977724516652d8cd47242546de587eac7870ab7585c6c0f7" exitCode=0 Nov 22 05:11:45 crc kubenswrapper[4818]: I1122 05:11:45.726546 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" event={"ID":"a527f913-2977-419c-9a4c-25b4c1729025","Type":"ContainerDied","Data":"a85c8dbbfe43d1f5977724516652d8cd47242546de587eac7870ab7585c6c0f7"} Nov 22 05:11:45 crc kubenswrapper[4818]: I1122 05:11:45.726698 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:11:45 crc kubenswrapper[4818]: I1122 05:11:45.726711 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" event={"ID":"a527f913-2977-419c-9a4c-25b4c1729025","Type":"ContainerStarted","Data":"dd79d8485fd85c4d0789a9fcddee8a153cd2a022fc0dce7e1cadc5e4a9cbaa05"} Nov 22 05:11:45 crc kubenswrapper[4818]: I1122 05:11:45.728721 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6c695f78-88f9-4013-a67b-4148203ff826","Type":"ContainerStarted","Data":"e80c78808446d40769f3b994d25731cc2980be59283e0ffed0567c28df6d861a"} Nov 22 05:11:46 crc kubenswrapper[4818]: I1122 05:11:46.582927 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:46 crc kubenswrapper[4818]: I1122 05:11:46.602834 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" podStartSLOduration=4.602815587 podStartE2EDuration="4.602815587s" podCreationTimestamp="2025-11-22 05:11:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:11:45.751153901 +0000 UTC m=+1458.325570448" watchObservedRunningTime="2025-11-22 05:11:46.602815587 +0000 UTC m=+1459.177232114" Nov 22 05:11:46 crc kubenswrapper[4818]: I1122 05:11:46.658820 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:46 crc kubenswrapper[4818]: I1122 05:11:46.742518 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6c695f78-88f9-4013-a67b-4148203ff826","Type":"ContainerStarted","Data":"79b5c7f5348564cf2e4dc3fa2d6c23dbcf8b7ee107909a43a2542ca677c976a9"} Nov 22 05:11:46 crc kubenswrapper[4818]: I1122 05:11:46.745803 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="23d760b6-0b81-4e56-855b-c54f8db80db7" containerName="cinder-api-log" containerID="cri-o://8a93f7f1f40f37da2a06bca4fd23a35c988571ff1dbe202979fbc17c43d3116a" gracePeriod=30 Nov 22 05:11:46 crc kubenswrapper[4818]: I1122 05:11:46.746005 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"23d760b6-0b81-4e56-855b-c54f8db80db7","Type":"ContainerStarted","Data":"4de6042ff44438efadb4b89f93cd86597ed01917ed837196ffcfeed712781c15"} Nov 22 05:11:46 crc kubenswrapper[4818]: I1122 05:11:46.746515 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 22 05:11:46 crc kubenswrapper[4818]: I1122 05:11:46.746551 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="23d760b6-0b81-4e56-855b-c54f8db80db7" containerName="cinder-api" containerID="cri-o://4de6042ff44438efadb4b89f93cd86597ed01917ed837196ffcfeed712781c15" gracePeriod=30 Nov 22 05:11:46 crc kubenswrapper[4818]: I1122 05:11:46.769941 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.851441385 podStartE2EDuration="4.769922664s" podCreationTimestamp="2025-11-22 05:11:42 +0000 UTC" firstStartedPulling="2025-11-22 05:11:43.871851848 +0000 UTC m=+1456.446268375" lastFinishedPulling="2025-11-22 05:11:44.790333127 +0000 UTC m=+1457.364749654" observedRunningTime="2025-11-22 05:11:46.763343481 +0000 UTC m=+1459.337760008" watchObservedRunningTime="2025-11-22 05:11:46.769922664 +0000 UTC m=+1459.344339191" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.509130 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.599365 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-combined-ca-bundle\") pod \"23d760b6-0b81-4e56-855b-c54f8db80db7\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.599420 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-config-data-custom\") pod \"23d760b6-0b81-4e56-855b-c54f8db80db7\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.599458 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-config-data\") pod \"23d760b6-0b81-4e56-855b-c54f8db80db7\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.599530 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jd49\" (UniqueName: \"kubernetes.io/projected/23d760b6-0b81-4e56-855b-c54f8db80db7-kube-api-access-4jd49\") pod \"23d760b6-0b81-4e56-855b-c54f8db80db7\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.599578 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d760b6-0b81-4e56-855b-c54f8db80db7-logs\") pod \"23d760b6-0b81-4e56-855b-c54f8db80db7\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.599599 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23d760b6-0b81-4e56-855b-c54f8db80db7-etc-machine-id\") pod \"23d760b6-0b81-4e56-855b-c54f8db80db7\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.599649 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-scripts\") pod \"23d760b6-0b81-4e56-855b-c54f8db80db7\" (UID: \"23d760b6-0b81-4e56-855b-c54f8db80db7\") " Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.600372 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/23d760b6-0b81-4e56-855b-c54f8db80db7-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "23d760b6-0b81-4e56-855b-c54f8db80db7" (UID: "23d760b6-0b81-4e56-855b-c54f8db80db7"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.600514 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23d760b6-0b81-4e56-855b-c54f8db80db7-logs" (OuterVolumeSpecName: "logs") pod "23d760b6-0b81-4e56-855b-c54f8db80db7" (UID: "23d760b6-0b81-4e56-855b-c54f8db80db7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.604901 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "23d760b6-0b81-4e56-855b-c54f8db80db7" (UID: "23d760b6-0b81-4e56-855b-c54f8db80db7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.605390 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23d760b6-0b81-4e56-855b-c54f8db80db7-kube-api-access-4jd49" (OuterVolumeSpecName: "kube-api-access-4jd49") pod "23d760b6-0b81-4e56-855b-c54f8db80db7" (UID: "23d760b6-0b81-4e56-855b-c54f8db80db7"). InnerVolumeSpecName "kube-api-access-4jd49". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.606366 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-scripts" (OuterVolumeSpecName: "scripts") pod "23d760b6-0b81-4e56-855b-c54f8db80db7" (UID: "23d760b6-0b81-4e56-855b-c54f8db80db7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.630069 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "23d760b6-0b81-4e56-855b-c54f8db80db7" (UID: "23d760b6-0b81-4e56-855b-c54f8db80db7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.685792 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-config-data" (OuterVolumeSpecName: "config-data") pod "23d760b6-0b81-4e56-855b-c54f8db80db7" (UID: "23d760b6-0b81-4e56-855b-c54f8db80db7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.701230 4818 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23d760b6-0b81-4e56-855b-c54f8db80db7-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.701282 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.701294 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.701306 4818 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.701316 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d760b6-0b81-4e56-855b-c54f8db80db7-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.701326 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jd49\" (UniqueName: \"kubernetes.io/projected/23d760b6-0b81-4e56-855b-c54f8db80db7-kube-api-access-4jd49\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.701339 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d760b6-0b81-4e56-855b-c54f8db80db7-logs\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.755751 4818 generic.go:334] "Generic (PLEG): container finished" podID="23d760b6-0b81-4e56-855b-c54f8db80db7" containerID="4de6042ff44438efadb4b89f93cd86597ed01917ed837196ffcfeed712781c15" exitCode=0 Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.755787 4818 generic.go:334] "Generic (PLEG): container finished" podID="23d760b6-0b81-4e56-855b-c54f8db80db7" containerID="8a93f7f1f40f37da2a06bca4fd23a35c988571ff1dbe202979fbc17c43d3116a" exitCode=143 Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.755869 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"23d760b6-0b81-4e56-855b-c54f8db80db7","Type":"ContainerDied","Data":"4de6042ff44438efadb4b89f93cd86597ed01917ed837196ffcfeed712781c15"} Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.755935 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"23d760b6-0b81-4e56-855b-c54f8db80db7","Type":"ContainerDied","Data":"8a93f7f1f40f37da2a06bca4fd23a35c988571ff1dbe202979fbc17c43d3116a"} Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.755954 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"23d760b6-0b81-4e56-855b-c54f8db80db7","Type":"ContainerDied","Data":"82e6291f33c21e6384845395ceed938e4b26b04ef9a0b053a6c776d419d81ece"} Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.755993 4818 scope.go:117] "RemoveContainer" containerID="4de6042ff44438efadb4b89f93cd86597ed01917ed837196ffcfeed712781c15" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.756696 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.779706 4818 scope.go:117] "RemoveContainer" containerID="8a93f7f1f40f37da2a06bca4fd23a35c988571ff1dbe202979fbc17c43d3116a" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.792410 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.799012 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.825179 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 22 05:11:47 crc kubenswrapper[4818]: E1122 05:11:47.825549 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64f9d89e-e211-4ee2-b78f-4321d612a468" containerName="dnsmasq-dns" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.825567 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="64f9d89e-e211-4ee2-b78f-4321d612a468" containerName="dnsmasq-dns" Nov 22 05:11:47 crc kubenswrapper[4818]: E1122 05:11:47.825579 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23d760b6-0b81-4e56-855b-c54f8db80db7" containerName="cinder-api-log" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.825585 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="23d760b6-0b81-4e56-855b-c54f8db80db7" containerName="cinder-api-log" Nov 22 05:11:47 crc kubenswrapper[4818]: E1122 05:11:47.825600 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23d760b6-0b81-4e56-855b-c54f8db80db7" containerName="cinder-api" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.825606 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="23d760b6-0b81-4e56-855b-c54f8db80db7" containerName="cinder-api" Nov 22 05:11:47 crc kubenswrapper[4818]: E1122 05:11:47.825623 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64f9d89e-e211-4ee2-b78f-4321d612a468" containerName="init" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.825628 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="64f9d89e-e211-4ee2-b78f-4321d612a468" containerName="init" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.825801 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="23d760b6-0b81-4e56-855b-c54f8db80db7" containerName="cinder-api" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.825812 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="23d760b6-0b81-4e56-855b-c54f8db80db7" containerName="cinder-api-log" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.825824 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="64f9d89e-e211-4ee2-b78f-4321d612a468" containerName="dnsmasq-dns" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.826667 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.829061 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.829464 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.829584 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.845688 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.855116 4818 scope.go:117] "RemoveContainer" containerID="4de6042ff44438efadb4b89f93cd86597ed01917ed837196ffcfeed712781c15" Nov 22 05:11:47 crc kubenswrapper[4818]: E1122 05:11:47.859452 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4de6042ff44438efadb4b89f93cd86597ed01917ed837196ffcfeed712781c15\": container with ID starting with 4de6042ff44438efadb4b89f93cd86597ed01917ed837196ffcfeed712781c15 not found: ID does not exist" containerID="4de6042ff44438efadb4b89f93cd86597ed01917ed837196ffcfeed712781c15" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.859508 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4de6042ff44438efadb4b89f93cd86597ed01917ed837196ffcfeed712781c15"} err="failed to get container status \"4de6042ff44438efadb4b89f93cd86597ed01917ed837196ffcfeed712781c15\": rpc error: code = NotFound desc = could not find container \"4de6042ff44438efadb4b89f93cd86597ed01917ed837196ffcfeed712781c15\": container with ID starting with 4de6042ff44438efadb4b89f93cd86597ed01917ed837196ffcfeed712781c15 not found: ID does not exist" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.859541 4818 scope.go:117] "RemoveContainer" containerID="8a93f7f1f40f37da2a06bca4fd23a35c988571ff1dbe202979fbc17c43d3116a" Nov 22 05:11:47 crc kubenswrapper[4818]: E1122 05:11:47.860778 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a93f7f1f40f37da2a06bca4fd23a35c988571ff1dbe202979fbc17c43d3116a\": container with ID starting with 8a93f7f1f40f37da2a06bca4fd23a35c988571ff1dbe202979fbc17c43d3116a not found: ID does not exist" containerID="8a93f7f1f40f37da2a06bca4fd23a35c988571ff1dbe202979fbc17c43d3116a" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.860830 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a93f7f1f40f37da2a06bca4fd23a35c988571ff1dbe202979fbc17c43d3116a"} err="failed to get container status \"8a93f7f1f40f37da2a06bca4fd23a35c988571ff1dbe202979fbc17c43d3116a\": rpc error: code = NotFound desc = could not find container \"8a93f7f1f40f37da2a06bca4fd23a35c988571ff1dbe202979fbc17c43d3116a\": container with ID starting with 8a93f7f1f40f37da2a06bca4fd23a35c988571ff1dbe202979fbc17c43d3116a not found: ID does not exist" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.860857 4818 scope.go:117] "RemoveContainer" containerID="4de6042ff44438efadb4b89f93cd86597ed01917ed837196ffcfeed712781c15" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.870400 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4de6042ff44438efadb4b89f93cd86597ed01917ed837196ffcfeed712781c15"} err="failed to get container status \"4de6042ff44438efadb4b89f93cd86597ed01917ed837196ffcfeed712781c15\": rpc error: code = NotFound desc = could not find container \"4de6042ff44438efadb4b89f93cd86597ed01917ed837196ffcfeed712781c15\": container with ID starting with 4de6042ff44438efadb4b89f93cd86597ed01917ed837196ffcfeed712781c15 not found: ID does not exist" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.870449 4818 scope.go:117] "RemoveContainer" containerID="8a93f7f1f40f37da2a06bca4fd23a35c988571ff1dbe202979fbc17c43d3116a" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.876406 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a93f7f1f40f37da2a06bca4fd23a35c988571ff1dbe202979fbc17c43d3116a"} err="failed to get container status \"8a93f7f1f40f37da2a06bca4fd23a35c988571ff1dbe202979fbc17c43d3116a\": rpc error: code = NotFound desc = could not find container \"8a93f7f1f40f37da2a06bca4fd23a35c988571ff1dbe202979fbc17c43d3116a\": container with ID starting with 8a93f7f1f40f37da2a06bca4fd23a35c988571ff1dbe202979fbc17c43d3116a not found: ID does not exist" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.909873 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a4f90e5-006b-43db-b910-429b496a7c65-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.909934 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a4f90e5-006b-43db-b910-429b496a7c65-logs\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.909954 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29gz9\" (UniqueName: \"kubernetes.io/projected/9a4f90e5-006b-43db-b910-429b496a7c65-kube-api-access-29gz9\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.909985 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a4f90e5-006b-43db-b910-429b496a7c65-config-data-custom\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.910004 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a4f90e5-006b-43db-b910-429b496a7c65-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.910023 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a4f90e5-006b-43db-b910-429b496a7c65-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.910062 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a4f90e5-006b-43db-b910-429b496a7c65-config-data\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.910080 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a4f90e5-006b-43db-b910-429b496a7c65-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:47 crc kubenswrapper[4818]: I1122 05:11:47.910141 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a4f90e5-006b-43db-b910-429b496a7c65-scripts\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.011799 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a4f90e5-006b-43db-b910-429b496a7c65-scripts\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.011910 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a4f90e5-006b-43db-b910-429b496a7c65-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.011958 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29gz9\" (UniqueName: \"kubernetes.io/projected/9a4f90e5-006b-43db-b910-429b496a7c65-kube-api-access-29gz9\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.011986 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a4f90e5-006b-43db-b910-429b496a7c65-logs\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.012044 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a4f90e5-006b-43db-b910-429b496a7c65-config-data-custom\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.012076 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a4f90e5-006b-43db-b910-429b496a7c65-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.012101 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a4f90e5-006b-43db-b910-429b496a7c65-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.012178 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a4f90e5-006b-43db-b910-429b496a7c65-config-data\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.012212 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a4f90e5-006b-43db-b910-429b496a7c65-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.013614 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a4f90e5-006b-43db-b910-429b496a7c65-logs\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.016198 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a4f90e5-006b-43db-b910-429b496a7c65-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.020593 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a4f90e5-006b-43db-b910-429b496a7c65-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.022756 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a4f90e5-006b-43db-b910-429b496a7c65-scripts\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.033559 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a4f90e5-006b-43db-b910-429b496a7c65-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.034409 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a4f90e5-006b-43db-b910-429b496a7c65-config-data\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.034748 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a4f90e5-006b-43db-b910-429b496a7c65-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.039997 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29gz9\" (UniqueName: \"kubernetes.io/projected/9a4f90e5-006b-43db-b910-429b496a7c65-kube-api-access-29gz9\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.040175 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a4f90e5-006b-43db-b910-429b496a7c65-config-data-custom\") pod \"cinder-api-0\" (UID: \"9a4f90e5-006b-43db-b910-429b496a7c65\") " pod="openstack/cinder-api-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.155182 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.225270 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.304637 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23d760b6-0b81-4e56-855b-c54f8db80db7" path="/var/lib/kubelet/pods/23d760b6-0b81-4e56-855b-c54f8db80db7/volumes" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.448946 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.540891 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6dfb48dcf4-6ph2v" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.591142 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-57dfcbc748-z27b5"] Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.591365 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-57dfcbc748-z27b5" podUID="fc2feeaa-d88f-4114-9e61-ae0df78e123c" containerName="barbican-api-log" containerID="cri-o://bf53fb270cfa817e4ab79364a9d5d6b7a4303fa2679fdb75110f922d09189d08" gracePeriod=30 Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.591721 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-57dfcbc748-z27b5" podUID="fc2feeaa-d88f-4114-9e61-ae0df78e123c" containerName="barbican-api" containerID="cri-o://61982d5f0de0f9899cc92f39d18fb52198b8bf6b87fdba3371b4d2182c507c44" gracePeriod=30 Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.600686 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7b464767cb-jcwc8" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.616204 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-57dfcbc748-z27b5" podUID="fc2feeaa-d88f-4114-9e61-ae0df78e123c" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.147:9311/healthcheck\": EOF" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.622619 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-57dfcbc748-z27b5" podUID="fc2feeaa-d88f-4114-9e61-ae0df78e123c" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.147:9311/healthcheck\": EOF" Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.641807 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.774212 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9a4f90e5-006b-43db-b910-429b496a7c65","Type":"ContainerStarted","Data":"9726c0c5f07fcf7c64d2406eeffd6d1058601afbb78eda496d598baf178ce2d0"} Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.779138 4818 generic.go:334] "Generic (PLEG): container finished" podID="fc2feeaa-d88f-4114-9e61-ae0df78e123c" containerID="bf53fb270cfa817e4ab79364a9d5d6b7a4303fa2679fdb75110f922d09189d08" exitCode=143 Nov 22 05:11:48 crc kubenswrapper[4818]: I1122 05:11:48.779201 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57dfcbc748-z27b5" event={"ID":"fc2feeaa-d88f-4114-9e61-ae0df78e123c","Type":"ContainerDied","Data":"bf53fb270cfa817e4ab79364a9d5d6b7a4303fa2679fdb75110f922d09189d08"} Nov 22 05:11:49 crc kubenswrapper[4818]: I1122 05:11:49.324467 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-54d7c687fd-74zvh" Nov 22 05:11:50 crc kubenswrapper[4818]: I1122 05:11:50.589074 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-ld8qn" podUID="4bf0b266-290e-4c4f-8ddb-17c38155751e" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 05:11:51 crc kubenswrapper[4818]: I1122 05:11:51.362658 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 22 05:11:51 crc kubenswrapper[4818]: I1122 05:11:51.368372 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 22 05:11:51 crc kubenswrapper[4818]: I1122 05:11:51.370362 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 22 05:11:51 crc kubenswrapper[4818]: I1122 05:11:51.370848 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-rx9kr" Nov 22 05:11:51 crc kubenswrapper[4818]: I1122 05:11:51.371752 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 22 05:11:51 crc kubenswrapper[4818]: I1122 05:11:51.385856 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 22 05:11:51 crc kubenswrapper[4818]: I1122 05:11:51.463618 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac27ad40-d8e9-4ebb-9623-3136a1f9753c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ac27ad40-d8e9-4ebb-9623-3136a1f9753c\") " pod="openstack/openstackclient" Nov 22 05:11:51 crc kubenswrapper[4818]: I1122 05:11:51.463729 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ac27ad40-d8e9-4ebb-9623-3136a1f9753c-openstack-config\") pod \"openstackclient\" (UID: \"ac27ad40-d8e9-4ebb-9623-3136a1f9753c\") " pod="openstack/openstackclient" Nov 22 05:11:51 crc kubenswrapper[4818]: I1122 05:11:51.463856 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mndq\" (UniqueName: \"kubernetes.io/projected/ac27ad40-d8e9-4ebb-9623-3136a1f9753c-kube-api-access-6mndq\") pod \"openstackclient\" (UID: \"ac27ad40-d8e9-4ebb-9623-3136a1f9753c\") " pod="openstack/openstackclient" Nov 22 05:11:51 crc kubenswrapper[4818]: I1122 05:11:51.463879 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ac27ad40-d8e9-4ebb-9623-3136a1f9753c-openstack-config-secret\") pod \"openstackclient\" (UID: \"ac27ad40-d8e9-4ebb-9623-3136a1f9753c\") " pod="openstack/openstackclient" Nov 22 05:11:51 crc kubenswrapper[4818]: I1122 05:11:51.565779 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac27ad40-d8e9-4ebb-9623-3136a1f9753c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ac27ad40-d8e9-4ebb-9623-3136a1f9753c\") " pod="openstack/openstackclient" Nov 22 05:11:51 crc kubenswrapper[4818]: I1122 05:11:51.565854 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ac27ad40-d8e9-4ebb-9623-3136a1f9753c-openstack-config\") pod \"openstackclient\" (UID: \"ac27ad40-d8e9-4ebb-9623-3136a1f9753c\") " pod="openstack/openstackclient" Nov 22 05:11:51 crc kubenswrapper[4818]: I1122 05:11:51.565937 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mndq\" (UniqueName: \"kubernetes.io/projected/ac27ad40-d8e9-4ebb-9623-3136a1f9753c-kube-api-access-6mndq\") pod \"openstackclient\" (UID: \"ac27ad40-d8e9-4ebb-9623-3136a1f9753c\") " pod="openstack/openstackclient" Nov 22 05:11:51 crc kubenswrapper[4818]: I1122 05:11:51.565959 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ac27ad40-d8e9-4ebb-9623-3136a1f9753c-openstack-config-secret\") pod \"openstackclient\" (UID: \"ac27ad40-d8e9-4ebb-9623-3136a1f9753c\") " pod="openstack/openstackclient" Nov 22 05:11:51 crc kubenswrapper[4818]: I1122 05:11:51.566860 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ac27ad40-d8e9-4ebb-9623-3136a1f9753c-openstack-config\") pod \"openstackclient\" (UID: \"ac27ad40-d8e9-4ebb-9623-3136a1f9753c\") " pod="openstack/openstackclient" Nov 22 05:11:51 crc kubenswrapper[4818]: I1122 05:11:51.578176 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ac27ad40-d8e9-4ebb-9623-3136a1f9753c-openstack-config-secret\") pod \"openstackclient\" (UID: \"ac27ad40-d8e9-4ebb-9623-3136a1f9753c\") " pod="openstack/openstackclient" Nov 22 05:11:51 crc kubenswrapper[4818]: I1122 05:11:51.581437 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac27ad40-d8e9-4ebb-9623-3136a1f9753c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ac27ad40-d8e9-4ebb-9623-3136a1f9753c\") " pod="openstack/openstackclient" Nov 22 05:11:51 crc kubenswrapper[4818]: I1122 05:11:51.584109 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mndq\" (UniqueName: \"kubernetes.io/projected/ac27ad40-d8e9-4ebb-9623-3136a1f9753c-kube-api-access-6mndq\") pod \"openstackclient\" (UID: \"ac27ad40-d8e9-4ebb-9623-3136a1f9753c\") " pod="openstack/openstackclient" Nov 22 05:11:51 crc kubenswrapper[4818]: I1122 05:11:51.642734 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9a4f90e5-006b-43db-b910-429b496a7c65","Type":"ContainerStarted","Data":"3db07f207aedaa2617084c95eb14663c5e5480167d435dcd3db9c422d79c62d8"} Nov 22 05:11:51 crc kubenswrapper[4818]: I1122 05:11:51.692804 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 22 05:11:52 crc kubenswrapper[4818]: W1122 05:11:52.146478 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac27ad40_d8e9_4ebb_9623_3136a1f9753c.slice/crio-8a9e957e0eff4af571b5d2a24171843e21fdf421fa48a03107166b073214dc3f WatchSource:0}: Error finding container 8a9e957e0eff4af571b5d2a24171843e21fdf421fa48a03107166b073214dc3f: Status 404 returned error can't find the container with id 8a9e957e0eff4af571b5d2a24171843e21fdf421fa48a03107166b073214dc3f Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.148149 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-57dfcbc748-z27b5" podUID="fc2feeaa-d88f-4114-9e61-ae0df78e123c" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.147:9311/healthcheck\": read tcp 10.217.0.2:56528->10.217.0.147:9311: read: connection reset by peer" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.148171 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-57dfcbc748-z27b5" podUID="fc2feeaa-d88f-4114-9e61-ae0df78e123c" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.147:9311/healthcheck\": read tcp 10.217.0.2:56518->10.217.0.147:9311: read: connection reset by peer" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.148604 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.566352 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.653005 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ac27ad40-d8e9-4ebb-9623-3136a1f9753c","Type":"ContainerStarted","Data":"8a9e957e0eff4af571b5d2a24171843e21fdf421fa48a03107166b073214dc3f"} Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.654991 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9a4f90e5-006b-43db-b910-429b496a7c65","Type":"ContainerStarted","Data":"201e8c54d27fc30b292685e7b276683bb993a55bbccb378db2ca905f3c6d21e3"} Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.656060 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.658017 4818 generic.go:334] "Generic (PLEG): container finished" podID="fc2feeaa-d88f-4114-9e61-ae0df78e123c" containerID="61982d5f0de0f9899cc92f39d18fb52198b8bf6b87fdba3371b4d2182c507c44" exitCode=0 Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.658061 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57dfcbc748-z27b5" event={"ID":"fc2feeaa-d88f-4114-9e61-ae0df78e123c","Type":"ContainerDied","Data":"61982d5f0de0f9899cc92f39d18fb52198b8bf6b87fdba3371b4d2182c507c44"} Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.658083 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57dfcbc748-z27b5" event={"ID":"fc2feeaa-d88f-4114-9e61-ae0df78e123c","Type":"ContainerDied","Data":"5d23ff50d49decdb3391c830a39d665163994acf9167d7c3da9b148f233b2c5e"} Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.658103 4818 scope.go:117] "RemoveContainer" containerID="61982d5f0de0f9899cc92f39d18fb52198b8bf6b87fdba3371b4d2182c507c44" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.658194 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57dfcbc748-z27b5" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.687012 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.686992265 podStartE2EDuration="5.686992265s" podCreationTimestamp="2025-11-22 05:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:11:52.675295976 +0000 UTC m=+1465.249712523" watchObservedRunningTime="2025-11-22 05:11:52.686992265 +0000 UTC m=+1465.261408792" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.688453 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhpkq\" (UniqueName: \"kubernetes.io/projected/fc2feeaa-d88f-4114-9e61-ae0df78e123c-kube-api-access-fhpkq\") pod \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\" (UID: \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\") " Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.688606 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc2feeaa-d88f-4114-9e61-ae0df78e123c-config-data\") pod \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\" (UID: \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\") " Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.688634 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc2feeaa-d88f-4114-9e61-ae0df78e123c-config-data-custom\") pod \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\" (UID: \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\") " Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.688656 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc2feeaa-d88f-4114-9e61-ae0df78e123c-logs\") pod \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\" (UID: \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\") " Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.688749 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc2feeaa-d88f-4114-9e61-ae0df78e123c-combined-ca-bundle\") pod \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\" (UID: \"fc2feeaa-d88f-4114-9e61-ae0df78e123c\") " Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.689210 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc2feeaa-d88f-4114-9e61-ae0df78e123c-logs" (OuterVolumeSpecName: "logs") pod "fc2feeaa-d88f-4114-9e61-ae0df78e123c" (UID: "fc2feeaa-d88f-4114-9e61-ae0df78e123c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.694105 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc2feeaa-d88f-4114-9e61-ae0df78e123c-kube-api-access-fhpkq" (OuterVolumeSpecName: "kube-api-access-fhpkq") pod "fc2feeaa-d88f-4114-9e61-ae0df78e123c" (UID: "fc2feeaa-d88f-4114-9e61-ae0df78e123c"). InnerVolumeSpecName "kube-api-access-fhpkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.694533 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc2feeaa-d88f-4114-9e61-ae0df78e123c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fc2feeaa-d88f-4114-9e61-ae0df78e123c" (UID: "fc2feeaa-d88f-4114-9e61-ae0df78e123c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.695142 4818 scope.go:117] "RemoveContainer" containerID="bf53fb270cfa817e4ab79364a9d5d6b7a4303fa2679fdb75110f922d09189d08" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.724180 4818 scope.go:117] "RemoveContainer" containerID="61982d5f0de0f9899cc92f39d18fb52198b8bf6b87fdba3371b4d2182c507c44" Nov 22 05:11:52 crc kubenswrapper[4818]: E1122 05:11:52.724600 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61982d5f0de0f9899cc92f39d18fb52198b8bf6b87fdba3371b4d2182c507c44\": container with ID starting with 61982d5f0de0f9899cc92f39d18fb52198b8bf6b87fdba3371b4d2182c507c44 not found: ID does not exist" containerID="61982d5f0de0f9899cc92f39d18fb52198b8bf6b87fdba3371b4d2182c507c44" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.724649 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61982d5f0de0f9899cc92f39d18fb52198b8bf6b87fdba3371b4d2182c507c44"} err="failed to get container status \"61982d5f0de0f9899cc92f39d18fb52198b8bf6b87fdba3371b4d2182c507c44\": rpc error: code = NotFound desc = could not find container \"61982d5f0de0f9899cc92f39d18fb52198b8bf6b87fdba3371b4d2182c507c44\": container with ID starting with 61982d5f0de0f9899cc92f39d18fb52198b8bf6b87fdba3371b4d2182c507c44 not found: ID does not exist" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.724717 4818 scope.go:117] "RemoveContainer" containerID="bf53fb270cfa817e4ab79364a9d5d6b7a4303fa2679fdb75110f922d09189d08" Nov 22 05:11:52 crc kubenswrapper[4818]: E1122 05:11:52.725060 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf53fb270cfa817e4ab79364a9d5d6b7a4303fa2679fdb75110f922d09189d08\": container with ID starting with bf53fb270cfa817e4ab79364a9d5d6b7a4303fa2679fdb75110f922d09189d08 not found: ID does not exist" containerID="bf53fb270cfa817e4ab79364a9d5d6b7a4303fa2679fdb75110f922d09189d08" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.725092 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf53fb270cfa817e4ab79364a9d5d6b7a4303fa2679fdb75110f922d09189d08"} err="failed to get container status \"bf53fb270cfa817e4ab79364a9d5d6b7a4303fa2679fdb75110f922d09189d08\": rpc error: code = NotFound desc = could not find container \"bf53fb270cfa817e4ab79364a9d5d6b7a4303fa2679fdb75110f922d09189d08\": container with ID starting with bf53fb270cfa817e4ab79364a9d5d6b7a4303fa2679fdb75110f922d09189d08 not found: ID does not exist" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.733231 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc2feeaa-d88f-4114-9e61-ae0df78e123c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc2feeaa-d88f-4114-9e61-ae0df78e123c" (UID: "fc2feeaa-d88f-4114-9e61-ae0df78e123c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.756444 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc2feeaa-d88f-4114-9e61-ae0df78e123c-config-data" (OuterVolumeSpecName: "config-data") pod "fc2feeaa-d88f-4114-9e61-ae0df78e123c" (UID: "fc2feeaa-d88f-4114-9e61-ae0df78e123c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.791769 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc2feeaa-d88f-4114-9e61-ae0df78e123c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.791804 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhpkq\" (UniqueName: \"kubernetes.io/projected/fc2feeaa-d88f-4114-9e61-ae0df78e123c-kube-api-access-fhpkq\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.791816 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc2feeaa-d88f-4114-9e61-ae0df78e123c-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.791824 4818 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc2feeaa-d88f-4114-9e61-ae0df78e123c-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.791833 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc2feeaa-d88f-4114-9e61-ae0df78e123c-logs\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.989923 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-57dfcbc748-z27b5"] Nov 22 05:11:52 crc kubenswrapper[4818]: I1122 05:11:52.996373 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-57dfcbc748-z27b5"] Nov 22 05:11:53 crc kubenswrapper[4818]: I1122 05:11:53.368351 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:11:53 crc kubenswrapper[4818]: I1122 05:11:53.442876 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-jlc6c"] Nov 22 05:11:53 crc kubenswrapper[4818]: I1122 05:11:53.444176 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" podUID="86b5909b-ca6a-4251-a03a-d0accd183c3a" containerName="dnsmasq-dns" containerID="cri-o://8b389f86c5110e71cbf7f20b4806d0a801afb4bd669d562a6283584396f1dd38" gracePeriod=10 Nov 22 05:11:53 crc kubenswrapper[4818]: I1122 05:11:53.479719 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 22 05:11:53 crc kubenswrapper[4818]: I1122 05:11:53.535130 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 05:11:53 crc kubenswrapper[4818]: I1122 05:11:53.676101 4818 generic.go:334] "Generic (PLEG): container finished" podID="86b5909b-ca6a-4251-a03a-d0accd183c3a" containerID="8b389f86c5110e71cbf7f20b4806d0a801afb4bd669d562a6283584396f1dd38" exitCode=0 Nov 22 05:11:53 crc kubenswrapper[4818]: I1122 05:11:53.676168 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" event={"ID":"86b5909b-ca6a-4251-a03a-d0accd183c3a","Type":"ContainerDied","Data":"8b389f86c5110e71cbf7f20b4806d0a801afb4bd669d562a6283584396f1dd38"} Nov 22 05:11:53 crc kubenswrapper[4818]: I1122 05:11:53.678500 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6c695f78-88f9-4013-a67b-4148203ff826" containerName="cinder-scheduler" containerID="cri-o://e80c78808446d40769f3b994d25731cc2980be59283e0ffed0567c28df6d861a" gracePeriod=30 Nov 22 05:11:53 crc kubenswrapper[4818]: I1122 05:11:53.679406 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6c695f78-88f9-4013-a67b-4148203ff826" containerName="probe" containerID="cri-o://79b5c7f5348564cf2e4dc3fa2d6c23dbcf8b7ee107909a43a2542ca677c976a9" gracePeriod=30 Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.000851 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.131343 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r99p2\" (UniqueName: \"kubernetes.io/projected/86b5909b-ca6a-4251-a03a-d0accd183c3a-kube-api-access-r99p2\") pod \"86b5909b-ca6a-4251-a03a-d0accd183c3a\" (UID: \"86b5909b-ca6a-4251-a03a-d0accd183c3a\") " Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.131409 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-ovsdbserver-sb\") pod \"86b5909b-ca6a-4251-a03a-d0accd183c3a\" (UID: \"86b5909b-ca6a-4251-a03a-d0accd183c3a\") " Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.131516 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-dns-svc\") pod \"86b5909b-ca6a-4251-a03a-d0accd183c3a\" (UID: \"86b5909b-ca6a-4251-a03a-d0accd183c3a\") " Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.131552 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-config\") pod \"86b5909b-ca6a-4251-a03a-d0accd183c3a\" (UID: \"86b5909b-ca6a-4251-a03a-d0accd183c3a\") " Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.131587 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-ovsdbserver-nb\") pod \"86b5909b-ca6a-4251-a03a-d0accd183c3a\" (UID: \"86b5909b-ca6a-4251-a03a-d0accd183c3a\") " Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.139383 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86b5909b-ca6a-4251-a03a-d0accd183c3a-kube-api-access-r99p2" (OuterVolumeSpecName: "kube-api-access-r99p2") pod "86b5909b-ca6a-4251-a03a-d0accd183c3a" (UID: "86b5909b-ca6a-4251-a03a-d0accd183c3a"). InnerVolumeSpecName "kube-api-access-r99p2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.184867 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-config" (OuterVolumeSpecName: "config") pod "86b5909b-ca6a-4251-a03a-d0accd183c3a" (UID: "86b5909b-ca6a-4251-a03a-d0accd183c3a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.185282 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "86b5909b-ca6a-4251-a03a-d0accd183c3a" (UID: "86b5909b-ca6a-4251-a03a-d0accd183c3a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.187760 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "86b5909b-ca6a-4251-a03a-d0accd183c3a" (UID: "86b5909b-ca6a-4251-a03a-d0accd183c3a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.208719 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "86b5909b-ca6a-4251-a03a-d0accd183c3a" (UID: "86b5909b-ca6a-4251-a03a-d0accd183c3a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.233769 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r99p2\" (UniqueName: \"kubernetes.io/projected/86b5909b-ca6a-4251-a03a-d0accd183c3a-kube-api-access-r99p2\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.233805 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.233829 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.233839 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.233846 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86b5909b-ca6a-4251-a03a-d0accd183c3a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.302114 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc2feeaa-d88f-4114-9e61-ae0df78e123c" path="/var/lib/kubelet/pods/fc2feeaa-d88f-4114-9e61-ae0df78e123c/volumes" Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.690677 4818 generic.go:334] "Generic (PLEG): container finished" podID="6c695f78-88f9-4013-a67b-4148203ff826" containerID="79b5c7f5348564cf2e4dc3fa2d6c23dbcf8b7ee107909a43a2542ca677c976a9" exitCode=0 Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.690748 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6c695f78-88f9-4013-a67b-4148203ff826","Type":"ContainerDied","Data":"79b5c7f5348564cf2e4dc3fa2d6c23dbcf8b7ee107909a43a2542ca677c976a9"} Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.693846 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" event={"ID":"86b5909b-ca6a-4251-a03a-d0accd183c3a","Type":"ContainerDied","Data":"5af9564cc8115832f57190aa122059843a317340f9f7480c200ba2db28712956"} Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.693910 4818 scope.go:117] "RemoveContainer" containerID="8b389f86c5110e71cbf7f20b4806d0a801afb4bd669d562a6283584396f1dd38" Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.693953 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-jlc6c" Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.738186 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-jlc6c"] Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.748028 4818 scope.go:117] "RemoveContainer" containerID="00b89d651bd7219dd8a3927dc06d45b661001ad95cba245901d5d110d9250041" Nov 22 05:11:54 crc kubenswrapper[4818]: I1122 05:11:54.748294 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-jlc6c"] Nov 22 05:11:56 crc kubenswrapper[4818]: I1122 05:11:56.304223 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86b5909b-ca6a-4251-a03a-d0accd183c3a" path="/var/lib/kubelet/pods/86b5909b-ca6a-4251-a03a-d0accd183c3a/volumes" Nov 22 05:11:56 crc kubenswrapper[4818]: I1122 05:11:56.726055 4818 generic.go:334] "Generic (PLEG): container finished" podID="6c695f78-88f9-4013-a67b-4148203ff826" containerID="e80c78808446d40769f3b994d25731cc2980be59283e0ffed0567c28df6d861a" exitCode=0 Nov 22 05:11:56 crc kubenswrapper[4818]: I1122 05:11:56.726105 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6c695f78-88f9-4013-a67b-4148203ff826","Type":"ContainerDied","Data":"e80c78808446d40769f3b994d25731cc2980be59283e0ffed0567c28df6d861a"} Nov 22 05:11:58 crc kubenswrapper[4818]: I1122 05:11:58.980528 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:11:59 crc kubenswrapper[4818]: I1122 05:11:59.972968 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:11:59 crc kubenswrapper[4818]: I1122 05:11:59.973968 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerName="proxy-httpd" containerID="cri-o://5a9ef5cee36644f0cb6cd57d370f33f472dc1aaed14c909eda40f0a18dc88eea" gracePeriod=30 Nov 22 05:11:59 crc kubenswrapper[4818]: I1122 05:11:59.974009 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerName="ceilometer-notification-agent" containerID="cri-o://4f4db4c8ecf36d0202f1a721660dec3efe45b3d18072d5cd524421af404e3daa" gracePeriod=30 Nov 22 05:11:59 crc kubenswrapper[4818]: I1122 05:11:59.974009 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerName="sg-core" containerID="cri-o://440f38a06792aeef09e5e5c5d8c5e80dee11e1a69f69f96e2daf060a4807dbf0" gracePeriod=30 Nov 22 05:11:59 crc kubenswrapper[4818]: I1122 05:11:59.973917 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerName="ceilometer-central-agent" containerID="cri-o://b522dea24eec23301001232d2f761e1aca84d0a3f33389e65a611df5534e9e3a" gracePeriod=30 Nov 22 05:11:59 crc kubenswrapper[4818]: I1122 05:11:59.986226 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 22 05:12:00 crc kubenswrapper[4818]: I1122 05:12:00.120834 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 22 05:12:00 crc kubenswrapper[4818]: I1122 05:12:00.777123 4818 generic.go:334] "Generic (PLEG): container finished" podID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerID="5a9ef5cee36644f0cb6cd57d370f33f472dc1aaed14c909eda40f0a18dc88eea" exitCode=0 Nov 22 05:12:00 crc kubenswrapper[4818]: I1122 05:12:00.777415 4818 generic.go:334] "Generic (PLEG): container finished" podID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerID="440f38a06792aeef09e5e5c5d8c5e80dee11e1a69f69f96e2daf060a4807dbf0" exitCode=2 Nov 22 05:12:00 crc kubenswrapper[4818]: I1122 05:12:00.777435 4818 generic.go:334] "Generic (PLEG): container finished" podID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerID="b522dea24eec23301001232d2f761e1aca84d0a3f33389e65a611df5534e9e3a" exitCode=0 Nov 22 05:12:00 crc kubenswrapper[4818]: I1122 05:12:00.777190 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c105a582-4a5b-4c1b-aac3-bbde151fcb1b","Type":"ContainerDied","Data":"5a9ef5cee36644f0cb6cd57d370f33f472dc1aaed14c909eda40f0a18dc88eea"} Nov 22 05:12:00 crc kubenswrapper[4818]: I1122 05:12:00.777468 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c105a582-4a5b-4c1b-aac3-bbde151fcb1b","Type":"ContainerDied","Data":"440f38a06792aeef09e5e5c5d8c5e80dee11e1a69f69f96e2daf060a4807dbf0"} Nov 22 05:12:00 crc kubenswrapper[4818]: I1122 05:12:00.777480 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c105a582-4a5b-4c1b-aac3-bbde151fcb1b","Type":"ContainerDied","Data":"b522dea24eec23301001232d2f761e1aca84d0a3f33389e65a611df5534e9e3a"} Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.005919 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-64754b546f-kqpj9" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.068135 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-85b5b4f6cd-lp7b9"] Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.068365 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-85b5b4f6cd-lp7b9" podUID="5578cc92-9b08-4c31-a73f-9fcda619ee85" containerName="neutron-api" containerID="cri-o://8331b76f4dea87944ae11fc2b5fcb21a8047259b56bd8d8fc8d6e021f6c31446" gracePeriod=30 Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.068801 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-85b5b4f6cd-lp7b9" podUID="5578cc92-9b08-4c31-a73f-9fcda619ee85" containerName="neutron-httpd" containerID="cri-o://9d4df4d517e61088fca1643c98ccaae797e79aa61382a96b1d80a483e5add0d2" gracePeriod=30 Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.602877 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.699799 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-scripts\") pod \"6c695f78-88f9-4013-a67b-4148203ff826\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.699858 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c695f78-88f9-4013-a67b-4148203ff826-etc-machine-id\") pod \"6c695f78-88f9-4013-a67b-4148203ff826\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.699910 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-config-data-custom\") pod \"6c695f78-88f9-4013-a67b-4148203ff826\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.699932 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-combined-ca-bundle\") pod \"6c695f78-88f9-4013-a67b-4148203ff826\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.700012 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6c695f78-88f9-4013-a67b-4148203ff826-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6c695f78-88f9-4013-a67b-4148203ff826" (UID: "6c695f78-88f9-4013-a67b-4148203ff826"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.700052 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-config-data\") pod \"6c695f78-88f9-4013-a67b-4148203ff826\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.700088 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5pnt\" (UniqueName: \"kubernetes.io/projected/6c695f78-88f9-4013-a67b-4148203ff826-kube-api-access-h5pnt\") pod \"6c695f78-88f9-4013-a67b-4148203ff826\" (UID: \"6c695f78-88f9-4013-a67b-4148203ff826\") " Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.700601 4818 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c695f78-88f9-4013-a67b-4148203ff826-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.707452 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c695f78-88f9-4013-a67b-4148203ff826-kube-api-access-h5pnt" (OuterVolumeSpecName: "kube-api-access-h5pnt") pod "6c695f78-88f9-4013-a67b-4148203ff826" (UID: "6c695f78-88f9-4013-a67b-4148203ff826"). InnerVolumeSpecName "kube-api-access-h5pnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.708345 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-scripts" (OuterVolumeSpecName: "scripts") pod "6c695f78-88f9-4013-a67b-4148203ff826" (UID: "6c695f78-88f9-4013-a67b-4148203ff826"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.712797 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6c695f78-88f9-4013-a67b-4148203ff826" (UID: "6c695f78-88f9-4013-a67b-4148203ff826"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.758530 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c695f78-88f9-4013-a67b-4148203ff826" (UID: "6c695f78-88f9-4013-a67b-4148203ff826"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.788143 4818 generic.go:334] "Generic (PLEG): container finished" podID="5578cc92-9b08-4c31-a73f-9fcda619ee85" containerID="9d4df4d517e61088fca1643c98ccaae797e79aa61382a96b1d80a483e5add0d2" exitCode=0 Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.788181 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85b5b4f6cd-lp7b9" event={"ID":"5578cc92-9b08-4c31-a73f-9fcda619ee85","Type":"ContainerDied","Data":"9d4df4d517e61088fca1643c98ccaae797e79aa61382a96b1d80a483e5add0d2"} Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.795036 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ac27ad40-d8e9-4ebb-9623-3136a1f9753c","Type":"ContainerStarted","Data":"8a1c931eaf53e2a8cc4d0cd7d94655f9f47496e9ebee4e1d6fc77a311f6626f4"} Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.798743 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6c695f78-88f9-4013-a67b-4148203ff826","Type":"ContainerDied","Data":"4359e0f2ce46891b93b796ed533d131540d570be39d27a3872e2b9000ce14ef5"} Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.798807 4818 scope.go:117] "RemoveContainer" containerID="79b5c7f5348564cf2e4dc3fa2d6c23dbcf8b7ee107909a43a2542ca677c976a9" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.798930 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-config-data" (OuterVolumeSpecName: "config-data") pod "6c695f78-88f9-4013-a67b-4148203ff826" (UID: "6c695f78-88f9-4013-a67b-4148203ff826"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.798966 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.803012 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.803055 4818 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.803071 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.803083 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c695f78-88f9-4013-a67b-4148203ff826-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.803095 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5pnt\" (UniqueName: \"kubernetes.io/projected/6c695f78-88f9-4013-a67b-4148203ff826-kube-api-access-h5pnt\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.824655 4818 scope.go:117] "RemoveContainer" containerID="e80c78808446d40769f3b994d25731cc2980be59283e0ffed0567c28df6d861a" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.840363 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.708120651 podStartE2EDuration="10.840338589s" podCreationTimestamp="2025-11-22 05:11:51 +0000 UTC" firstStartedPulling="2025-11-22 05:11:52.14866786 +0000 UTC m=+1464.723084387" lastFinishedPulling="2025-11-22 05:12:01.280885808 +0000 UTC m=+1473.855302325" observedRunningTime="2025-11-22 05:12:01.818816352 +0000 UTC m=+1474.393232899" watchObservedRunningTime="2025-11-22 05:12:01.840338589 +0000 UTC m=+1474.414755126" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.849390 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.862041 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.865907 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.143:3000/\": dial tcp 10.217.0.143:3000: connect: connection refused" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.876852 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 05:12:01 crc kubenswrapper[4818]: E1122 05:12:01.877392 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c695f78-88f9-4013-a67b-4148203ff826" containerName="probe" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.877417 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c695f78-88f9-4013-a67b-4148203ff826" containerName="probe" Nov 22 05:12:01 crc kubenswrapper[4818]: E1122 05:12:01.877440 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc2feeaa-d88f-4114-9e61-ae0df78e123c" containerName="barbican-api-log" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.877449 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc2feeaa-d88f-4114-9e61-ae0df78e123c" containerName="barbican-api-log" Nov 22 05:12:01 crc kubenswrapper[4818]: E1122 05:12:01.877461 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86b5909b-ca6a-4251-a03a-d0accd183c3a" containerName="init" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.877469 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="86b5909b-ca6a-4251-a03a-d0accd183c3a" containerName="init" Nov 22 05:12:01 crc kubenswrapper[4818]: E1122 05:12:01.877489 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86b5909b-ca6a-4251-a03a-d0accd183c3a" containerName="dnsmasq-dns" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.877497 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="86b5909b-ca6a-4251-a03a-d0accd183c3a" containerName="dnsmasq-dns" Nov 22 05:12:01 crc kubenswrapper[4818]: E1122 05:12:01.877515 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc2feeaa-d88f-4114-9e61-ae0df78e123c" containerName="barbican-api" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.877524 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc2feeaa-d88f-4114-9e61-ae0df78e123c" containerName="barbican-api" Nov 22 05:12:01 crc kubenswrapper[4818]: E1122 05:12:01.877546 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c695f78-88f9-4013-a67b-4148203ff826" containerName="cinder-scheduler" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.877555 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c695f78-88f9-4013-a67b-4148203ff826" containerName="cinder-scheduler" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.877757 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c695f78-88f9-4013-a67b-4148203ff826" containerName="probe" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.877775 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c695f78-88f9-4013-a67b-4148203ff826" containerName="cinder-scheduler" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.877789 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc2feeaa-d88f-4114-9e61-ae0df78e123c" containerName="barbican-api" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.877800 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="86b5909b-ca6a-4251-a03a-d0accd183c3a" containerName="dnsmasq-dns" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.877813 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc2feeaa-d88f-4114-9e61-ae0df78e123c" containerName="barbican-api-log" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.878879 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.881355 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 22 05:12:01 crc kubenswrapper[4818]: I1122 05:12:01.902505 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.006898 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0fce1ca4-c9be-4e2a-947a-3cc652edeb50-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0fce1ca4-c9be-4e2a-947a-3cc652edeb50\") " pod="openstack/cinder-scheduler-0" Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.006996 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fce1ca4-c9be-4e2a-947a-3cc652edeb50-scripts\") pod \"cinder-scheduler-0\" (UID: \"0fce1ca4-c9be-4e2a-947a-3cc652edeb50\") " pod="openstack/cinder-scheduler-0" Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.007261 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0fce1ca4-c9be-4e2a-947a-3cc652edeb50-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0fce1ca4-c9be-4e2a-947a-3cc652edeb50\") " pod="openstack/cinder-scheduler-0" Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.007382 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fce1ca4-c9be-4e2a-947a-3cc652edeb50-config-data\") pod \"cinder-scheduler-0\" (UID: \"0fce1ca4-c9be-4e2a-947a-3cc652edeb50\") " pod="openstack/cinder-scheduler-0" Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.007442 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv9v7\" (UniqueName: \"kubernetes.io/projected/0fce1ca4-c9be-4e2a-947a-3cc652edeb50-kube-api-access-lv9v7\") pod \"cinder-scheduler-0\" (UID: \"0fce1ca4-c9be-4e2a-947a-3cc652edeb50\") " pod="openstack/cinder-scheduler-0" Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.007606 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fce1ca4-c9be-4e2a-947a-3cc652edeb50-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0fce1ca4-c9be-4e2a-947a-3cc652edeb50\") " pod="openstack/cinder-scheduler-0" Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.109104 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fce1ca4-c9be-4e2a-947a-3cc652edeb50-config-data\") pod \"cinder-scheduler-0\" (UID: \"0fce1ca4-c9be-4e2a-947a-3cc652edeb50\") " pod="openstack/cinder-scheduler-0" Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.109174 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv9v7\" (UniqueName: \"kubernetes.io/projected/0fce1ca4-c9be-4e2a-947a-3cc652edeb50-kube-api-access-lv9v7\") pod \"cinder-scheduler-0\" (UID: \"0fce1ca4-c9be-4e2a-947a-3cc652edeb50\") " pod="openstack/cinder-scheduler-0" Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.109222 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fce1ca4-c9be-4e2a-947a-3cc652edeb50-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0fce1ca4-c9be-4e2a-947a-3cc652edeb50\") " pod="openstack/cinder-scheduler-0" Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.109269 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0fce1ca4-c9be-4e2a-947a-3cc652edeb50-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0fce1ca4-c9be-4e2a-947a-3cc652edeb50\") " pod="openstack/cinder-scheduler-0" Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.109356 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fce1ca4-c9be-4e2a-947a-3cc652edeb50-scripts\") pod \"cinder-scheduler-0\" (UID: \"0fce1ca4-c9be-4e2a-947a-3cc652edeb50\") " pod="openstack/cinder-scheduler-0" Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.109440 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0fce1ca4-c9be-4e2a-947a-3cc652edeb50-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0fce1ca4-c9be-4e2a-947a-3cc652edeb50\") " pod="openstack/cinder-scheduler-0" Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.109534 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0fce1ca4-c9be-4e2a-947a-3cc652edeb50-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0fce1ca4-c9be-4e2a-947a-3cc652edeb50\") " pod="openstack/cinder-scheduler-0" Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.114000 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fce1ca4-c9be-4e2a-947a-3cc652edeb50-scripts\") pod \"cinder-scheduler-0\" (UID: \"0fce1ca4-c9be-4e2a-947a-3cc652edeb50\") " pod="openstack/cinder-scheduler-0" Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.115903 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fce1ca4-c9be-4e2a-947a-3cc652edeb50-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0fce1ca4-c9be-4e2a-947a-3cc652edeb50\") " pod="openstack/cinder-scheduler-0" Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.116227 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fce1ca4-c9be-4e2a-947a-3cc652edeb50-config-data\") pod \"cinder-scheduler-0\" (UID: \"0fce1ca4-c9be-4e2a-947a-3cc652edeb50\") " pod="openstack/cinder-scheduler-0" Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.116309 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0fce1ca4-c9be-4e2a-947a-3cc652edeb50-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0fce1ca4-c9be-4e2a-947a-3cc652edeb50\") " pod="openstack/cinder-scheduler-0" Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.130915 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv9v7\" (UniqueName: \"kubernetes.io/projected/0fce1ca4-c9be-4e2a-947a-3cc652edeb50-kube-api-access-lv9v7\") pod \"cinder-scheduler-0\" (UID: \"0fce1ca4-c9be-4e2a-947a-3cc652edeb50\") " pod="openstack/cinder-scheduler-0" Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.200514 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.304187 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c695f78-88f9-4013-a67b-4148203ff826" path="/var/lib/kubelet/pods/6c695f78-88f9-4013-a67b-4148203ff826/volumes" Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.761853 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 05:12:02 crc kubenswrapper[4818]: I1122 05:12:02.809557 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0fce1ca4-c9be-4e2a-947a-3cc652edeb50","Type":"ContainerStarted","Data":"3639f30c2165397f9a08e97437602d7a35f683a59646798982bbf230cffe4e2e"} Nov 22 05:12:03 crc kubenswrapper[4818]: I1122 05:12:03.811569 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 05:12:03 crc kubenswrapper[4818]: I1122 05:12:03.812186 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="92a5743a-8fbe-49a7-a653-2b32e6034f52" containerName="kube-state-metrics" containerID="cri-o://705def928e920052d7ab5b829dc7e1b212eeb3dcaf5b1e7ed12249a5873c659b" gracePeriod=30 Nov 22 05:12:03 crc kubenswrapper[4818]: I1122 05:12:03.842297 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:12:03 crc kubenswrapper[4818]: I1122 05:12:03.866509 4818 generic.go:334] "Generic (PLEG): container finished" podID="5578cc92-9b08-4c31-a73f-9fcda619ee85" containerID="8331b76f4dea87944ae11fc2b5fcb21a8047259b56bd8d8fc8d6e021f6c31446" exitCode=0 Nov 22 05:12:03 crc kubenswrapper[4818]: I1122 05:12:03.866609 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85b5b4f6cd-lp7b9" event={"ID":"5578cc92-9b08-4c31-a73f-9fcda619ee85","Type":"ContainerDied","Data":"8331b76f4dea87944ae11fc2b5fcb21a8047259b56bd8d8fc8d6e021f6c31446"} Nov 22 05:12:03 crc kubenswrapper[4818]: I1122 05:12:03.866635 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85b5b4f6cd-lp7b9" event={"ID":"5578cc92-9b08-4c31-a73f-9fcda619ee85","Type":"ContainerDied","Data":"06e597820204a4d2bf0d67108be5daf4600ea2278275ec786036526dfd5c64e1"} Nov 22 05:12:03 crc kubenswrapper[4818]: I1122 05:12:03.866652 4818 scope.go:117] "RemoveContainer" containerID="9d4df4d517e61088fca1643c98ccaae797e79aa61382a96b1d80a483e5add0d2" Nov 22 05:12:03 crc kubenswrapper[4818]: I1122 05:12:03.883623 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0fce1ca4-c9be-4e2a-947a-3cc652edeb50","Type":"ContainerStarted","Data":"42a8b50a25af2bf479273cc6f21d8db14d83606aa9e11fbaaef9a6d4b9c8e945"} Nov 22 05:12:03 crc kubenswrapper[4818]: I1122 05:12:03.948951 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-config\") pod \"5578cc92-9b08-4c31-a73f-9fcda619ee85\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " Nov 22 05:12:03 crc kubenswrapper[4818]: I1122 05:12:03.949025 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-combined-ca-bundle\") pod \"5578cc92-9b08-4c31-a73f-9fcda619ee85\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " Nov 22 05:12:03 crc kubenswrapper[4818]: I1122 05:12:03.949103 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvx4z\" (UniqueName: \"kubernetes.io/projected/5578cc92-9b08-4c31-a73f-9fcda619ee85-kube-api-access-jvx4z\") pod \"5578cc92-9b08-4c31-a73f-9fcda619ee85\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " Nov 22 05:12:03 crc kubenswrapper[4818]: I1122 05:12:03.949227 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-ovndb-tls-certs\") pod \"5578cc92-9b08-4c31-a73f-9fcda619ee85\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " Nov 22 05:12:03 crc kubenswrapper[4818]: I1122 05:12:03.949271 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-httpd-config\") pod \"5578cc92-9b08-4c31-a73f-9fcda619ee85\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " Nov 22 05:12:03 crc kubenswrapper[4818]: I1122 05:12:03.960475 4818 scope.go:117] "RemoveContainer" containerID="8331b76f4dea87944ae11fc2b5fcb21a8047259b56bd8d8fc8d6e021f6c31446" Nov 22 05:12:03 crc kubenswrapper[4818]: I1122 05:12:03.965481 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "5578cc92-9b08-4c31-a73f-9fcda619ee85" (UID: "5578cc92-9b08-4c31-a73f-9fcda619ee85"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:03 crc kubenswrapper[4818]: I1122 05:12:03.971404 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5578cc92-9b08-4c31-a73f-9fcda619ee85-kube-api-access-jvx4z" (OuterVolumeSpecName: "kube-api-access-jvx4z") pod "5578cc92-9b08-4c31-a73f-9fcda619ee85" (UID: "5578cc92-9b08-4c31-a73f-9fcda619ee85"). InnerVolumeSpecName "kube-api-access-jvx4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.020454 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-config" (OuterVolumeSpecName: "config") pod "5578cc92-9b08-4c31-a73f-9fcda619ee85" (UID: "5578cc92-9b08-4c31-a73f-9fcda619ee85"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.050356 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5578cc92-9b08-4c31-a73f-9fcda619ee85" (UID: "5578cc92-9b08-4c31-a73f-9fcda619ee85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.050532 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-combined-ca-bundle\") pod \"5578cc92-9b08-4c31-a73f-9fcda619ee85\" (UID: \"5578cc92-9b08-4c31-a73f-9fcda619ee85\") " Nov 22 05:12:04 crc kubenswrapper[4818]: W1122 05:12:04.050626 4818 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/5578cc92-9b08-4c31-a73f-9fcda619ee85/volumes/kubernetes.io~secret/combined-ca-bundle Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.050643 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5578cc92-9b08-4c31-a73f-9fcda619ee85" (UID: "5578cc92-9b08-4c31-a73f-9fcda619ee85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.051093 4818 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.051122 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.051135 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.051149 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvx4z\" (UniqueName: \"kubernetes.io/projected/5578cc92-9b08-4c31-a73f-9fcda619ee85-kube-api-access-jvx4z\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.060465 4818 scope.go:117] "RemoveContainer" containerID="9d4df4d517e61088fca1643c98ccaae797e79aa61382a96b1d80a483e5add0d2" Nov 22 05:12:04 crc kubenswrapper[4818]: E1122 05:12:04.061069 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d4df4d517e61088fca1643c98ccaae797e79aa61382a96b1d80a483e5add0d2\": container with ID starting with 9d4df4d517e61088fca1643c98ccaae797e79aa61382a96b1d80a483e5add0d2 not found: ID does not exist" containerID="9d4df4d517e61088fca1643c98ccaae797e79aa61382a96b1d80a483e5add0d2" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.061105 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d4df4d517e61088fca1643c98ccaae797e79aa61382a96b1d80a483e5add0d2"} err="failed to get container status \"9d4df4d517e61088fca1643c98ccaae797e79aa61382a96b1d80a483e5add0d2\": rpc error: code = NotFound desc = could not find container \"9d4df4d517e61088fca1643c98ccaae797e79aa61382a96b1d80a483e5add0d2\": container with ID starting with 9d4df4d517e61088fca1643c98ccaae797e79aa61382a96b1d80a483e5add0d2 not found: ID does not exist" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.061134 4818 scope.go:117] "RemoveContainer" containerID="8331b76f4dea87944ae11fc2b5fcb21a8047259b56bd8d8fc8d6e021f6c31446" Nov 22 05:12:04 crc kubenswrapper[4818]: E1122 05:12:04.061653 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8331b76f4dea87944ae11fc2b5fcb21a8047259b56bd8d8fc8d6e021f6c31446\": container with ID starting with 8331b76f4dea87944ae11fc2b5fcb21a8047259b56bd8d8fc8d6e021f6c31446 not found: ID does not exist" containerID="8331b76f4dea87944ae11fc2b5fcb21a8047259b56bd8d8fc8d6e021f6c31446" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.061680 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8331b76f4dea87944ae11fc2b5fcb21a8047259b56bd8d8fc8d6e021f6c31446"} err="failed to get container status \"8331b76f4dea87944ae11fc2b5fcb21a8047259b56bd8d8fc8d6e021f6c31446\": rpc error: code = NotFound desc = could not find container \"8331b76f4dea87944ae11fc2b5fcb21a8047259b56bd8d8fc8d6e021f6c31446\": container with ID starting with 8331b76f4dea87944ae11fc2b5fcb21a8047259b56bd8d8fc8d6e021f6c31446 not found: ID does not exist" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.086540 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "5578cc92-9b08-4c31-a73f-9fcda619ee85" (UID: "5578cc92-9b08-4c31-a73f-9fcda619ee85"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.153929 4818 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5578cc92-9b08-4c31-a73f-9fcda619ee85-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.442782 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.560836 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld5wh\" (UniqueName: \"kubernetes.io/projected/92a5743a-8fbe-49a7-a653-2b32e6034f52-kube-api-access-ld5wh\") pod \"92a5743a-8fbe-49a7-a653-2b32e6034f52\" (UID: \"92a5743a-8fbe-49a7-a653-2b32e6034f52\") " Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.574384 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92a5743a-8fbe-49a7-a653-2b32e6034f52-kube-api-access-ld5wh" (OuterVolumeSpecName: "kube-api-access-ld5wh") pod "92a5743a-8fbe-49a7-a653-2b32e6034f52" (UID: "92a5743a-8fbe-49a7-a653-2b32e6034f52"). InnerVolumeSpecName "kube-api-access-ld5wh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.662458 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld5wh\" (UniqueName: \"kubernetes.io/projected/92a5743a-8fbe-49a7-a653-2b32e6034f52-kube-api-access-ld5wh\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.892282 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85b5b4f6cd-lp7b9" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.894835 4818 generic.go:334] "Generic (PLEG): container finished" podID="92a5743a-8fbe-49a7-a653-2b32e6034f52" containerID="705def928e920052d7ab5b829dc7e1b212eeb3dcaf5b1e7ed12249a5873c659b" exitCode=2 Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.894882 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.894933 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"92a5743a-8fbe-49a7-a653-2b32e6034f52","Type":"ContainerDied","Data":"705def928e920052d7ab5b829dc7e1b212eeb3dcaf5b1e7ed12249a5873c659b"} Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.894963 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"92a5743a-8fbe-49a7-a653-2b32e6034f52","Type":"ContainerDied","Data":"8188a18c53b1e138d4155e4ec3e290ac468eb920c4360d5f90bbf0a1bf6e1b7c"} Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.894981 4818 scope.go:117] "RemoveContainer" containerID="705def928e920052d7ab5b829dc7e1b212eeb3dcaf5b1e7ed12249a5873c659b" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.897542 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0fce1ca4-c9be-4e2a-947a-3cc652edeb50","Type":"ContainerStarted","Data":"442a5d5e5c20d1c8500e957cc8ac264192eed23b1ccdcb537297de8a9434f741"} Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.911570 4818 scope.go:117] "RemoveContainer" containerID="705def928e920052d7ab5b829dc7e1b212eeb3dcaf5b1e7ed12249a5873c659b" Nov 22 05:12:04 crc kubenswrapper[4818]: E1122 05:12:04.911948 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"705def928e920052d7ab5b829dc7e1b212eeb3dcaf5b1e7ed12249a5873c659b\": container with ID starting with 705def928e920052d7ab5b829dc7e1b212eeb3dcaf5b1e7ed12249a5873c659b not found: ID does not exist" containerID="705def928e920052d7ab5b829dc7e1b212eeb3dcaf5b1e7ed12249a5873c659b" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.911986 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"705def928e920052d7ab5b829dc7e1b212eeb3dcaf5b1e7ed12249a5873c659b"} err="failed to get container status \"705def928e920052d7ab5b829dc7e1b212eeb3dcaf5b1e7ed12249a5873c659b\": rpc error: code = NotFound desc = could not find container \"705def928e920052d7ab5b829dc7e1b212eeb3dcaf5b1e7ed12249a5873c659b\": container with ID starting with 705def928e920052d7ab5b829dc7e1b212eeb3dcaf5b1e7ed12249a5873c659b not found: ID does not exist" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.920788 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-85b5b4f6cd-lp7b9"] Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.933010 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-85b5b4f6cd-lp7b9"] Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.945793 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.945772044 podStartE2EDuration="3.945772044s" podCreationTimestamp="2025-11-22 05:12:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:12:04.940226337 +0000 UTC m=+1477.514642864" watchObservedRunningTime="2025-11-22 05:12:04.945772044 +0000 UTC m=+1477.520188561" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.966242 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.976816 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.985296 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 05:12:04 crc kubenswrapper[4818]: E1122 05:12:04.985641 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a5743a-8fbe-49a7-a653-2b32e6034f52" containerName="kube-state-metrics" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.985658 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a5743a-8fbe-49a7-a653-2b32e6034f52" containerName="kube-state-metrics" Nov 22 05:12:04 crc kubenswrapper[4818]: E1122 05:12:04.985672 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5578cc92-9b08-4c31-a73f-9fcda619ee85" containerName="neutron-httpd" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.985679 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="5578cc92-9b08-4c31-a73f-9fcda619ee85" containerName="neutron-httpd" Nov 22 05:12:04 crc kubenswrapper[4818]: E1122 05:12:04.985706 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5578cc92-9b08-4c31-a73f-9fcda619ee85" containerName="neutron-api" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.985713 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="5578cc92-9b08-4c31-a73f-9fcda619ee85" containerName="neutron-api" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.985867 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="5578cc92-9b08-4c31-a73f-9fcda619ee85" containerName="neutron-api" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.985894 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="5578cc92-9b08-4c31-a73f-9fcda619ee85" containerName="neutron-httpd" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.985906 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="92a5743a-8fbe-49a7-a653-2b32e6034f52" containerName="kube-state-metrics" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.986441 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.988520 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.988661 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 22 05:12:04 crc kubenswrapper[4818]: I1122 05:12:04.998454 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.070030 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd030ac8-aebe-4cde-b0e8-403f3049a922-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"fd030ac8-aebe-4cde-b0e8-403f3049a922\") " pod="openstack/kube-state-metrics-0" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.070193 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd030ac8-aebe-4cde-b0e8-403f3049a922-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"fd030ac8-aebe-4cde-b0e8-403f3049a922\") " pod="openstack/kube-state-metrics-0" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.070230 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd68f\" (UniqueName: \"kubernetes.io/projected/fd030ac8-aebe-4cde-b0e8-403f3049a922-kube-api-access-sd68f\") pod \"kube-state-metrics-0\" (UID: \"fd030ac8-aebe-4cde-b0e8-403f3049a922\") " pod="openstack/kube-state-metrics-0" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.070247 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/fd030ac8-aebe-4cde-b0e8-403f3049a922-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"fd030ac8-aebe-4cde-b0e8-403f3049a922\") " pod="openstack/kube-state-metrics-0" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.171772 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd030ac8-aebe-4cde-b0e8-403f3049a922-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"fd030ac8-aebe-4cde-b0e8-403f3049a922\") " pod="openstack/kube-state-metrics-0" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.171841 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd030ac8-aebe-4cde-b0e8-403f3049a922-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"fd030ac8-aebe-4cde-b0e8-403f3049a922\") " pod="openstack/kube-state-metrics-0" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.171875 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd68f\" (UniqueName: \"kubernetes.io/projected/fd030ac8-aebe-4cde-b0e8-403f3049a922-kube-api-access-sd68f\") pod \"kube-state-metrics-0\" (UID: \"fd030ac8-aebe-4cde-b0e8-403f3049a922\") " pod="openstack/kube-state-metrics-0" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.171894 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/fd030ac8-aebe-4cde-b0e8-403f3049a922-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"fd030ac8-aebe-4cde-b0e8-403f3049a922\") " pod="openstack/kube-state-metrics-0" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.176908 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/fd030ac8-aebe-4cde-b0e8-403f3049a922-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"fd030ac8-aebe-4cde-b0e8-403f3049a922\") " pod="openstack/kube-state-metrics-0" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.177820 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd030ac8-aebe-4cde-b0e8-403f3049a922-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"fd030ac8-aebe-4cde-b0e8-403f3049a922\") " pod="openstack/kube-state-metrics-0" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.183196 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd030ac8-aebe-4cde-b0e8-403f3049a922-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"fd030ac8-aebe-4cde-b0e8-403f3049a922\") " pod="openstack/kube-state-metrics-0" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.192114 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd68f\" (UniqueName: \"kubernetes.io/projected/fd030ac8-aebe-4cde-b0e8-403f3049a922-kube-api-access-sd68f\") pod \"kube-state-metrics-0\" (UID: \"fd030ac8-aebe-4cde-b0e8-403f3049a922\") " pod="openstack/kube-state-metrics-0" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.303642 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.630773 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.781875 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-sg-core-conf-yaml\") pod \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.782345 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-log-httpd\") pod \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.782382 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-scripts\") pod \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.782414 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvsjb\" (UniqueName: \"kubernetes.io/projected/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-kube-api-access-vvsjb\") pod \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.782450 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-run-httpd\") pod \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.782472 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-combined-ca-bundle\") pod \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.782541 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-config-data\") pod \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\" (UID: \"c105a582-4a5b-4c1b-aac3-bbde151fcb1b\") " Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.782869 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c105a582-4a5b-4c1b-aac3-bbde151fcb1b" (UID: "c105a582-4a5b-4c1b-aac3-bbde151fcb1b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.782917 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c105a582-4a5b-4c1b-aac3-bbde151fcb1b" (UID: "c105a582-4a5b-4c1b-aac3-bbde151fcb1b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.783001 4818 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.783016 4818 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.790691 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-kube-api-access-vvsjb" (OuterVolumeSpecName: "kube-api-access-vvsjb") pod "c105a582-4a5b-4c1b-aac3-bbde151fcb1b" (UID: "c105a582-4a5b-4c1b-aac3-bbde151fcb1b"). InnerVolumeSpecName "kube-api-access-vvsjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.790871 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-scripts" (OuterVolumeSpecName: "scripts") pod "c105a582-4a5b-4c1b-aac3-bbde151fcb1b" (UID: "c105a582-4a5b-4c1b-aac3-bbde151fcb1b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.803433 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.837456 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c105a582-4a5b-4c1b-aac3-bbde151fcb1b" (UID: "c105a582-4a5b-4c1b-aac3-bbde151fcb1b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.878367 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c105a582-4a5b-4c1b-aac3-bbde151fcb1b" (UID: "c105a582-4a5b-4c1b-aac3-bbde151fcb1b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.885408 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.885455 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvsjb\" (UniqueName: \"kubernetes.io/projected/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-kube-api-access-vvsjb\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.885470 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.885481 4818 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.917443 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"fd030ac8-aebe-4cde-b0e8-403f3049a922","Type":"ContainerStarted","Data":"078035553a142aa777bdfbe2e2e3e5d9508e3552672f242ca838bc585cb64c9c"} Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.924894 4818 generic.go:334] "Generic (PLEG): container finished" podID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerID="4f4db4c8ecf36d0202f1a721660dec3efe45b3d18072d5cd524421af404e3daa" exitCode=0 Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.924964 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c105a582-4a5b-4c1b-aac3-bbde151fcb1b","Type":"ContainerDied","Data":"4f4db4c8ecf36d0202f1a721660dec3efe45b3d18072d5cd524421af404e3daa"} Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.924991 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c105a582-4a5b-4c1b-aac3-bbde151fcb1b","Type":"ContainerDied","Data":"913bea3087c6389e694663ed264cd7da7ca3c53ba4fbd657f836e3897e944fc5"} Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.925007 4818 scope.go:117] "RemoveContainer" containerID="5a9ef5cee36644f0cb6cd57d370f33f472dc1aaed14c909eda40f0a18dc88eea" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.925111 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.925195 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-config-data" (OuterVolumeSpecName: "config-data") pod "c105a582-4a5b-4c1b-aac3-bbde151fcb1b" (UID: "c105a582-4a5b-4c1b-aac3-bbde151fcb1b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.950717 4818 scope.go:117] "RemoveContainer" containerID="440f38a06792aeef09e5e5c5d8c5e80dee11e1a69f69f96e2daf060a4807dbf0" Nov 22 05:12:05 crc kubenswrapper[4818]: I1122 05:12:05.988915 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c105a582-4a5b-4c1b-aac3-bbde151fcb1b-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.004319 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.005643 4818 scope.go:117] "RemoveContainer" containerID="4f4db4c8ecf36d0202f1a721660dec3efe45b3d18072d5cd524421af404e3daa" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.029562 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.041497 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:06 crc kubenswrapper[4818]: E1122 05:12:06.041922 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerName="proxy-httpd" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.041935 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerName="proxy-httpd" Nov 22 05:12:06 crc kubenswrapper[4818]: E1122 05:12:06.041945 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerName="ceilometer-notification-agent" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.041952 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerName="ceilometer-notification-agent" Nov 22 05:12:06 crc kubenswrapper[4818]: E1122 05:12:06.041962 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerName="sg-core" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.041969 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerName="sg-core" Nov 22 05:12:06 crc kubenswrapper[4818]: E1122 05:12:06.041985 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerName="ceilometer-central-agent" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.041991 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerName="ceilometer-central-agent" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.042146 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerName="ceilometer-central-agent" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.042155 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerName="ceilometer-notification-agent" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.042167 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerName="sg-core" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.042178 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" containerName="proxy-httpd" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.044050 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.046203 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.046852 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.047110 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.049467 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.068367 4818 scope.go:117] "RemoveContainer" containerID="b522dea24eec23301001232d2f761e1aca84d0a3f33389e65a611df5534e9e3a" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.099553 4818 scope.go:117] "RemoveContainer" containerID="5a9ef5cee36644f0cb6cd57d370f33f472dc1aaed14c909eda40f0a18dc88eea" Nov 22 05:12:06 crc kubenswrapper[4818]: E1122 05:12:06.100117 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a9ef5cee36644f0cb6cd57d370f33f472dc1aaed14c909eda40f0a18dc88eea\": container with ID starting with 5a9ef5cee36644f0cb6cd57d370f33f472dc1aaed14c909eda40f0a18dc88eea not found: ID does not exist" containerID="5a9ef5cee36644f0cb6cd57d370f33f472dc1aaed14c909eda40f0a18dc88eea" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.100145 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a9ef5cee36644f0cb6cd57d370f33f472dc1aaed14c909eda40f0a18dc88eea"} err="failed to get container status \"5a9ef5cee36644f0cb6cd57d370f33f472dc1aaed14c909eda40f0a18dc88eea\": rpc error: code = NotFound desc = could not find container \"5a9ef5cee36644f0cb6cd57d370f33f472dc1aaed14c909eda40f0a18dc88eea\": container with ID starting with 5a9ef5cee36644f0cb6cd57d370f33f472dc1aaed14c909eda40f0a18dc88eea not found: ID does not exist" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.100167 4818 scope.go:117] "RemoveContainer" containerID="440f38a06792aeef09e5e5c5d8c5e80dee11e1a69f69f96e2daf060a4807dbf0" Nov 22 05:12:06 crc kubenswrapper[4818]: E1122 05:12:06.100534 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"440f38a06792aeef09e5e5c5d8c5e80dee11e1a69f69f96e2daf060a4807dbf0\": container with ID starting with 440f38a06792aeef09e5e5c5d8c5e80dee11e1a69f69f96e2daf060a4807dbf0 not found: ID does not exist" containerID="440f38a06792aeef09e5e5c5d8c5e80dee11e1a69f69f96e2daf060a4807dbf0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.100586 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"440f38a06792aeef09e5e5c5d8c5e80dee11e1a69f69f96e2daf060a4807dbf0"} err="failed to get container status \"440f38a06792aeef09e5e5c5d8c5e80dee11e1a69f69f96e2daf060a4807dbf0\": rpc error: code = NotFound desc = could not find container \"440f38a06792aeef09e5e5c5d8c5e80dee11e1a69f69f96e2daf060a4807dbf0\": container with ID starting with 440f38a06792aeef09e5e5c5d8c5e80dee11e1a69f69f96e2daf060a4807dbf0 not found: ID does not exist" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.100611 4818 scope.go:117] "RemoveContainer" containerID="4f4db4c8ecf36d0202f1a721660dec3efe45b3d18072d5cd524421af404e3daa" Nov 22 05:12:06 crc kubenswrapper[4818]: E1122 05:12:06.102614 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f4db4c8ecf36d0202f1a721660dec3efe45b3d18072d5cd524421af404e3daa\": container with ID starting with 4f4db4c8ecf36d0202f1a721660dec3efe45b3d18072d5cd524421af404e3daa not found: ID does not exist" containerID="4f4db4c8ecf36d0202f1a721660dec3efe45b3d18072d5cd524421af404e3daa" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.102634 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f4db4c8ecf36d0202f1a721660dec3efe45b3d18072d5cd524421af404e3daa"} err="failed to get container status \"4f4db4c8ecf36d0202f1a721660dec3efe45b3d18072d5cd524421af404e3daa\": rpc error: code = NotFound desc = could not find container \"4f4db4c8ecf36d0202f1a721660dec3efe45b3d18072d5cd524421af404e3daa\": container with ID starting with 4f4db4c8ecf36d0202f1a721660dec3efe45b3d18072d5cd524421af404e3daa not found: ID does not exist" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.102648 4818 scope.go:117] "RemoveContainer" containerID="b522dea24eec23301001232d2f761e1aca84d0a3f33389e65a611df5534e9e3a" Nov 22 05:12:06 crc kubenswrapper[4818]: E1122 05:12:06.102883 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b522dea24eec23301001232d2f761e1aca84d0a3f33389e65a611df5534e9e3a\": container with ID starting with b522dea24eec23301001232d2f761e1aca84d0a3f33389e65a611df5534e9e3a not found: ID does not exist" containerID="b522dea24eec23301001232d2f761e1aca84d0a3f33389e65a611df5534e9e3a" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.102906 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b522dea24eec23301001232d2f761e1aca84d0a3f33389e65a611df5534e9e3a"} err="failed to get container status \"b522dea24eec23301001232d2f761e1aca84d0a3f33389e65a611df5534e9e3a\": rpc error: code = NotFound desc = could not find container \"b522dea24eec23301001232d2f761e1aca84d0a3f33389e65a611df5534e9e3a\": container with ID starting with b522dea24eec23301001232d2f761e1aca84d0a3f33389e65a611df5534e9e3a not found: ID does not exist" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.193511 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.193573 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.193596 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-scripts\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.193662 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e7681b7-8bcf-4273-a20d-8e5a758589f1-run-httpd\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.193687 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e7681b7-8bcf-4273-a20d-8e5a758589f1-log-httpd\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.193752 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7jtr\" (UniqueName: \"kubernetes.io/projected/0e7681b7-8bcf-4273-a20d-8e5a758589f1-kube-api-access-g7jtr\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.193774 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.193802 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-config-data\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.295191 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7jtr\" (UniqueName: \"kubernetes.io/projected/0e7681b7-8bcf-4273-a20d-8e5a758589f1-kube-api-access-g7jtr\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.295877 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.296001 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-config-data\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.296136 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.296279 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.296385 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-scripts\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.296557 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e7681b7-8bcf-4273-a20d-8e5a758589f1-run-httpd\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.296669 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e7681b7-8bcf-4273-a20d-8e5a758589f1-log-httpd\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.297225 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e7681b7-8bcf-4273-a20d-8e5a758589f1-run-httpd\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.297280 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e7681b7-8bcf-4273-a20d-8e5a758589f1-log-httpd\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.300343 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-scripts\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.301651 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.304133 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.307199 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5578cc92-9b08-4c31-a73f-9fcda619ee85" path="/var/lib/kubelet/pods/5578cc92-9b08-4c31-a73f-9fcda619ee85/volumes" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.310497 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92a5743a-8fbe-49a7-a653-2b32e6034f52" path="/var/lib/kubelet/pods/92a5743a-8fbe-49a7-a653-2b32e6034f52/volumes" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.311047 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c105a582-4a5b-4c1b-aac3-bbde151fcb1b" path="/var/lib/kubelet/pods/c105a582-4a5b-4c1b-aac3-bbde151fcb1b/volumes" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.314267 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-config-data\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.314548 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.334870 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:06 crc kubenswrapper[4818]: E1122 05:12:06.335617 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-g7jtr], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="0e7681b7-8bcf-4273-a20d-8e5a758589f1" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.337647 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7jtr\" (UniqueName: \"kubernetes.io/projected/0e7681b7-8bcf-4273-a20d-8e5a758589f1-kube-api-access-g7jtr\") pod \"ceilometer-0\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.937281 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"fd030ac8-aebe-4cde-b0e8-403f3049a922","Type":"ContainerStarted","Data":"dc99ed6acbc4a4c87ef52e62353099a2f4f5e2dc7f5afbaa390d9987087a8c70"} Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.937651 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.938411 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.951154 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:12:06 crc kubenswrapper[4818]: I1122 05:12:06.957344 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.572688413 podStartE2EDuration="2.957331355s" podCreationTimestamp="2025-11-22 05:12:04 +0000 UTC" firstStartedPulling="2025-11-22 05:12:05.808688087 +0000 UTC m=+1478.383104614" lastFinishedPulling="2025-11-22 05:12:06.193331029 +0000 UTC m=+1478.767747556" observedRunningTime="2025-11-22 05:12:06.954316295 +0000 UTC m=+1479.528732822" watchObservedRunningTime="2025-11-22 05:12:06.957331355 +0000 UTC m=+1479.531747882" Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.108742 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-combined-ca-bundle\") pod \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.108972 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e7681b7-8bcf-4273-a20d-8e5a758589f1-log-httpd\") pod \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.109011 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7jtr\" (UniqueName: \"kubernetes.io/projected/0e7681b7-8bcf-4273-a20d-8e5a758589f1-kube-api-access-g7jtr\") pod \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.109068 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-config-data\") pod \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.109108 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e7681b7-8bcf-4273-a20d-8e5a758589f1-run-httpd\") pod \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.109532 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-ceilometer-tls-certs\") pod \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.109284 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e7681b7-8bcf-4273-a20d-8e5a758589f1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0e7681b7-8bcf-4273-a20d-8e5a758589f1" (UID: "0e7681b7-8bcf-4273-a20d-8e5a758589f1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.109472 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e7681b7-8bcf-4273-a20d-8e5a758589f1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0e7681b7-8bcf-4273-a20d-8e5a758589f1" (UID: "0e7681b7-8bcf-4273-a20d-8e5a758589f1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.109615 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-scripts\") pod \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.109977 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-sg-core-conf-yaml\") pod \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\" (UID: \"0e7681b7-8bcf-4273-a20d-8e5a758589f1\") " Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.110573 4818 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e7681b7-8bcf-4273-a20d-8e5a758589f1-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.110589 4818 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e7681b7-8bcf-4273-a20d-8e5a758589f1-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.114549 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e7681b7-8bcf-4273-a20d-8e5a758589f1-kube-api-access-g7jtr" (OuterVolumeSpecName: "kube-api-access-g7jtr") pod "0e7681b7-8bcf-4273-a20d-8e5a758589f1" (UID: "0e7681b7-8bcf-4273-a20d-8e5a758589f1"). InnerVolumeSpecName "kube-api-access-g7jtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.114564 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0e7681b7-8bcf-4273-a20d-8e5a758589f1" (UID: "0e7681b7-8bcf-4273-a20d-8e5a758589f1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.114543 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-config-data" (OuterVolumeSpecName: "config-data") pod "0e7681b7-8bcf-4273-a20d-8e5a758589f1" (UID: "0e7681b7-8bcf-4273-a20d-8e5a758589f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.114731 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e7681b7-8bcf-4273-a20d-8e5a758589f1" (UID: "0e7681b7-8bcf-4273-a20d-8e5a758589f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.116221 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "0e7681b7-8bcf-4273-a20d-8e5a758589f1" (UID: "0e7681b7-8bcf-4273-a20d-8e5a758589f1"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.122424 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-scripts" (OuterVolumeSpecName: "scripts") pod "0e7681b7-8bcf-4273-a20d-8e5a758589f1" (UID: "0e7681b7-8bcf-4273-a20d-8e5a758589f1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.201394 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.212170 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7jtr\" (UniqueName: \"kubernetes.io/projected/0e7681b7-8bcf-4273-a20d-8e5a758589f1-kube-api-access-g7jtr\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.212316 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.212330 4818 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.212342 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.212354 4818 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.212366 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e7681b7-8bcf-4273-a20d-8e5a758589f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.946791 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:12:07 crc kubenswrapper[4818]: I1122 05:12:07.994238 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.002241 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.020454 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.022762 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.024604 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.024756 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.024913 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.090282 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.129610 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-config-data\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.129653 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/38c10a66-aae4-4e0e-839d-aa86e7d619c9-run-httpd\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.129682 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.129703 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/38c10a66-aae4-4e0e-839d-aa86e7d619c9-log-httpd\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.129742 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2jz7\" (UniqueName: \"kubernetes.io/projected/38c10a66-aae4-4e0e-839d-aa86e7d619c9-kube-api-access-j2jz7\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.129759 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-scripts\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.129778 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.129940 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.231740 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-scripts\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.232066 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2jz7\" (UniqueName: \"kubernetes.io/projected/38c10a66-aae4-4e0e-839d-aa86e7d619c9-kube-api-access-j2jz7\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.232092 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.232136 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.232243 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/38c10a66-aae4-4e0e-839d-aa86e7d619c9-run-httpd\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.232288 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-config-data\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.232325 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.232358 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/38c10a66-aae4-4e0e-839d-aa86e7d619c9-log-httpd\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.233017 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/38c10a66-aae4-4e0e-839d-aa86e7d619c9-log-httpd\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.233290 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/38c10a66-aae4-4e0e-839d-aa86e7d619c9-run-httpd\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.237511 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.237593 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.238181 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-scripts\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.238440 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-config-data\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.240637 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.262713 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2jz7\" (UniqueName: \"kubernetes.io/projected/38c10a66-aae4-4e0e-839d-aa86e7d619c9-kube-api-access-j2jz7\") pod \"ceilometer-0\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.302367 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e7681b7-8bcf-4273-a20d-8e5a758589f1" path="/var/lib/kubelet/pods/0e7681b7-8bcf-4273-a20d-8e5a758589f1/volumes" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.338830 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.813246 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:08 crc kubenswrapper[4818]: W1122 05:12:08.816547 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38c10a66_aae4_4e0e_839d_aa86e7d619c9.slice/crio-2768445f3c85d14c8f7211723d5aec5a43eca4bbbe27933089a82098b9142edf WatchSource:0}: Error finding container 2768445f3c85d14c8f7211723d5aec5a43eca4bbbe27933089a82098b9142edf: Status 404 returned error can't find the container with id 2768445f3c85d14c8f7211723d5aec5a43eca4bbbe27933089a82098b9142edf Nov 22 05:12:08 crc kubenswrapper[4818]: I1122 05:12:08.973102 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"38c10a66-aae4-4e0e-839d-aa86e7d619c9","Type":"ContainerStarted","Data":"2768445f3c85d14c8f7211723d5aec5a43eca4bbbe27933089a82098b9142edf"} Nov 22 05:12:09 crc kubenswrapper[4818]: I1122 05:12:09.471274 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:09 crc kubenswrapper[4818]: I1122 05:12:09.988766 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"38c10a66-aae4-4e0e-839d-aa86e7d619c9","Type":"ContainerStarted","Data":"e0c6e0e1acc116e5e9152826b2ef6dff3b73a7d824ed2fa9dba73782f1d203d6"} Nov 22 05:12:10 crc kubenswrapper[4818]: I1122 05:12:10.998994 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"38c10a66-aae4-4e0e-839d-aa86e7d619c9","Type":"ContainerStarted","Data":"adb1fc6502127770a258a3477ac90aaadb29d71775af8433af6d2f3a90f5cc31"} Nov 22 05:12:10 crc kubenswrapper[4818]: I1122 05:12:10.999290 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"38c10a66-aae4-4e0e-839d-aa86e7d619c9","Type":"ContainerStarted","Data":"3b6026ce62631ee3a487dfa65840eb6c1a3482ca6723b042a74ea5d32ab32697"} Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.351656 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-8vsln"] Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.353327 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8vsln" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.366052 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8vsln"] Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.474096 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-d92ms"] Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.475413 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-d92ms" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.503715 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-d92ms"] Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.504741 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/39310079-a73f-4fe1-854e-aacffbddcabb-operator-scripts\") pod \"nova-api-db-create-8vsln\" (UID: \"39310079-a73f-4fe1-854e-aacffbddcabb\") " pod="openstack/nova-api-db-create-8vsln" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.504874 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zqcm\" (UniqueName: \"kubernetes.io/projected/39310079-a73f-4fe1-854e-aacffbddcabb-kube-api-access-4zqcm\") pod \"nova-api-db-create-8vsln\" (UID: \"39310079-a73f-4fe1-854e-aacffbddcabb\") " pod="openstack/nova-api-db-create-8vsln" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.536051 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.560735 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-b75c-account-create-ztmdp"] Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.561881 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b75c-account-create-ztmdp" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.564127 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.586308 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-b75c-account-create-ztmdp"] Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.609233 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/39310079-a73f-4fe1-854e-aacffbddcabb-operator-scripts\") pod \"nova-api-db-create-8vsln\" (UID: \"39310079-a73f-4fe1-854e-aacffbddcabb\") " pod="openstack/nova-api-db-create-8vsln" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.609555 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79wmw\" (UniqueName: \"kubernetes.io/projected/057a60f1-d0da-4590-ab6c-4de6cf4c79d1-kube-api-access-79wmw\") pod \"nova-cell0-db-create-d92ms\" (UID: \"057a60f1-d0da-4590-ab6c-4de6cf4c79d1\") " pod="openstack/nova-cell0-db-create-d92ms" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.609623 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/057a60f1-d0da-4590-ab6c-4de6cf4c79d1-operator-scripts\") pod \"nova-cell0-db-create-d92ms\" (UID: \"057a60f1-d0da-4590-ab6c-4de6cf4c79d1\") " pod="openstack/nova-cell0-db-create-d92ms" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.609679 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zqcm\" (UniqueName: \"kubernetes.io/projected/39310079-a73f-4fe1-854e-aacffbddcabb-kube-api-access-4zqcm\") pod \"nova-api-db-create-8vsln\" (UID: \"39310079-a73f-4fe1-854e-aacffbddcabb\") " pod="openstack/nova-api-db-create-8vsln" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.611705 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/39310079-a73f-4fe1-854e-aacffbddcabb-operator-scripts\") pod \"nova-api-db-create-8vsln\" (UID: \"39310079-a73f-4fe1-854e-aacffbddcabb\") " pod="openstack/nova-api-db-create-8vsln" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.634514 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zqcm\" (UniqueName: \"kubernetes.io/projected/39310079-a73f-4fe1-854e-aacffbddcabb-kube-api-access-4zqcm\") pod \"nova-api-db-create-8vsln\" (UID: \"39310079-a73f-4fe1-854e-aacffbddcabb\") " pod="openstack/nova-api-db-create-8vsln" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.649010 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-5njh6"] Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.651958 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5njh6" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.658893 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-5njh6"] Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.692925 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8vsln" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.725544 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e0926ba-d1e1-470f-91a0-221471d29797-operator-scripts\") pod \"nova-api-b75c-account-create-ztmdp\" (UID: \"6e0926ba-d1e1-470f-91a0-221471d29797\") " pod="openstack/nova-api-b75c-account-create-ztmdp" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.725638 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2xgq\" (UniqueName: \"kubernetes.io/projected/6e0926ba-d1e1-470f-91a0-221471d29797-kube-api-access-l2xgq\") pod \"nova-api-b75c-account-create-ztmdp\" (UID: \"6e0926ba-d1e1-470f-91a0-221471d29797\") " pod="openstack/nova-api-b75c-account-create-ztmdp" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.725777 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79wmw\" (UniqueName: \"kubernetes.io/projected/057a60f1-d0da-4590-ab6c-4de6cf4c79d1-kube-api-access-79wmw\") pod \"nova-cell0-db-create-d92ms\" (UID: \"057a60f1-d0da-4590-ab6c-4de6cf4c79d1\") " pod="openstack/nova-cell0-db-create-d92ms" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.725854 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/057a60f1-d0da-4590-ab6c-4de6cf4c79d1-operator-scripts\") pod \"nova-cell0-db-create-d92ms\" (UID: \"057a60f1-d0da-4590-ab6c-4de6cf4c79d1\") " pod="openstack/nova-cell0-db-create-d92ms" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.726656 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/057a60f1-d0da-4590-ab6c-4de6cf4c79d1-operator-scripts\") pod \"nova-cell0-db-create-d92ms\" (UID: \"057a60f1-d0da-4590-ab6c-4de6cf4c79d1\") " pod="openstack/nova-cell0-db-create-d92ms" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.757784 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79wmw\" (UniqueName: \"kubernetes.io/projected/057a60f1-d0da-4590-ab6c-4de6cf4c79d1-kube-api-access-79wmw\") pod \"nova-cell0-db-create-d92ms\" (UID: \"057a60f1-d0da-4590-ab6c-4de6cf4c79d1\") " pod="openstack/nova-cell0-db-create-d92ms" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.784855 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-36d8-account-create-bwv78"] Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.786043 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-36d8-account-create-bwv78" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.789309 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.791130 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-36d8-account-create-bwv78"] Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.819866 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-d92ms" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.827658 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e0926ba-d1e1-470f-91a0-221471d29797-operator-scripts\") pod \"nova-api-b75c-account-create-ztmdp\" (UID: \"6e0926ba-d1e1-470f-91a0-221471d29797\") " pod="openstack/nova-api-b75c-account-create-ztmdp" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.827718 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjnqd\" (UniqueName: \"kubernetes.io/projected/5a04faa3-41d0-4e96-8957-097cba1b5154-kube-api-access-bjnqd\") pod \"nova-cell1-db-create-5njh6\" (UID: \"5a04faa3-41d0-4e96-8957-097cba1b5154\") " pod="openstack/nova-cell1-db-create-5njh6" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.827743 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2xgq\" (UniqueName: \"kubernetes.io/projected/6e0926ba-d1e1-470f-91a0-221471d29797-kube-api-access-l2xgq\") pod \"nova-api-b75c-account-create-ztmdp\" (UID: \"6e0926ba-d1e1-470f-91a0-221471d29797\") " pod="openstack/nova-api-b75c-account-create-ztmdp" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.827772 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a04faa3-41d0-4e96-8957-097cba1b5154-operator-scripts\") pod \"nova-cell1-db-create-5njh6\" (UID: \"5a04faa3-41d0-4e96-8957-097cba1b5154\") " pod="openstack/nova-cell1-db-create-5njh6" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.828632 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e0926ba-d1e1-470f-91a0-221471d29797-operator-scripts\") pod \"nova-api-b75c-account-create-ztmdp\" (UID: \"6e0926ba-d1e1-470f-91a0-221471d29797\") " pod="openstack/nova-api-b75c-account-create-ztmdp" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.850351 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2xgq\" (UniqueName: \"kubernetes.io/projected/6e0926ba-d1e1-470f-91a0-221471d29797-kube-api-access-l2xgq\") pod \"nova-api-b75c-account-create-ztmdp\" (UID: \"6e0926ba-d1e1-470f-91a0-221471d29797\") " pod="openstack/nova-api-b75c-account-create-ztmdp" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.880080 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b75c-account-create-ztmdp" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.930031 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5382656-d791-455a-b7a4-925c2b37d399-operator-scripts\") pod \"nova-cell0-36d8-account-create-bwv78\" (UID: \"f5382656-d791-455a-b7a4-925c2b37d399\") " pod="openstack/nova-cell0-36d8-account-create-bwv78" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.930632 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjnqd\" (UniqueName: \"kubernetes.io/projected/5a04faa3-41d0-4e96-8957-097cba1b5154-kube-api-access-bjnqd\") pod \"nova-cell1-db-create-5njh6\" (UID: \"5a04faa3-41d0-4e96-8957-097cba1b5154\") " pod="openstack/nova-cell1-db-create-5njh6" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.930688 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9jbx\" (UniqueName: \"kubernetes.io/projected/f5382656-d791-455a-b7a4-925c2b37d399-kube-api-access-f9jbx\") pod \"nova-cell0-36d8-account-create-bwv78\" (UID: \"f5382656-d791-455a-b7a4-925c2b37d399\") " pod="openstack/nova-cell0-36d8-account-create-bwv78" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.930718 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a04faa3-41d0-4e96-8957-097cba1b5154-operator-scripts\") pod \"nova-cell1-db-create-5njh6\" (UID: \"5a04faa3-41d0-4e96-8957-097cba1b5154\") " pod="openstack/nova-cell1-db-create-5njh6" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.936657 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a04faa3-41d0-4e96-8957-097cba1b5154-operator-scripts\") pod \"nova-cell1-db-create-5njh6\" (UID: \"5a04faa3-41d0-4e96-8957-097cba1b5154\") " pod="openstack/nova-cell1-db-create-5njh6" Nov 22 05:12:12 crc kubenswrapper[4818]: I1122 05:12:12.958001 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjnqd\" (UniqueName: \"kubernetes.io/projected/5a04faa3-41d0-4e96-8957-097cba1b5154-kube-api-access-bjnqd\") pod \"nova-cell1-db-create-5njh6\" (UID: \"5a04faa3-41d0-4e96-8957-097cba1b5154\") " pod="openstack/nova-cell1-db-create-5njh6" Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:12.999889 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cce9-account-create-qjwmx"] Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.001049 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cce9-account-create-qjwmx" Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.002903 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.006799 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5njh6" Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.028941 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cce9-account-create-qjwmx"] Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.032314 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9jbx\" (UniqueName: \"kubernetes.io/projected/f5382656-d791-455a-b7a4-925c2b37d399-kube-api-access-f9jbx\") pod \"nova-cell0-36d8-account-create-bwv78\" (UID: \"f5382656-d791-455a-b7a4-925c2b37d399\") " pod="openstack/nova-cell0-36d8-account-create-bwv78" Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.032368 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5382656-d791-455a-b7a4-925c2b37d399-operator-scripts\") pod \"nova-cell0-36d8-account-create-bwv78\" (UID: \"f5382656-d791-455a-b7a4-925c2b37d399\") " pod="openstack/nova-cell0-36d8-account-create-bwv78" Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.033065 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5382656-d791-455a-b7a4-925c2b37d399-operator-scripts\") pod \"nova-cell0-36d8-account-create-bwv78\" (UID: \"f5382656-d791-455a-b7a4-925c2b37d399\") " pod="openstack/nova-cell0-36d8-account-create-bwv78" Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.054470 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9jbx\" (UniqueName: \"kubernetes.io/projected/f5382656-d791-455a-b7a4-925c2b37d399-kube-api-access-f9jbx\") pod \"nova-cell0-36d8-account-create-bwv78\" (UID: \"f5382656-d791-455a-b7a4-925c2b37d399\") " pod="openstack/nova-cell0-36d8-account-create-bwv78" Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.070932 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"38c10a66-aae4-4e0e-839d-aa86e7d619c9","Type":"ContainerStarted","Data":"e5a594e737babf68be143838f05e73d933ece2f93df8f917e60d57e3fa0da9e9"} Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.071118 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" containerName="ceilometer-central-agent" containerID="cri-o://e0c6e0e1acc116e5e9152826b2ef6dff3b73a7d824ed2fa9dba73782f1d203d6" gracePeriod=30 Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.071384 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.071625 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" containerName="proxy-httpd" containerID="cri-o://e5a594e737babf68be143838f05e73d933ece2f93df8f917e60d57e3fa0da9e9" gracePeriod=30 Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.071678 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" containerName="sg-core" containerID="cri-o://adb1fc6502127770a258a3477ac90aaadb29d71775af8433af6d2f3a90f5cc31" gracePeriod=30 Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.071698 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" containerName="ceilometer-notification-agent" containerID="cri-o://3b6026ce62631ee3a487dfa65840eb6c1a3482ca6723b042a74ea5d32ab32697" gracePeriod=30 Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.114727 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-36d8-account-create-bwv78" Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.119109 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.770390999 podStartE2EDuration="5.119087568s" podCreationTimestamp="2025-11-22 05:12:08 +0000 UTC" firstStartedPulling="2025-11-22 05:12:08.81891961 +0000 UTC m=+1481.393336137" lastFinishedPulling="2025-11-22 05:12:12.167616159 +0000 UTC m=+1484.742032706" observedRunningTime="2025-11-22 05:12:13.104518013 +0000 UTC m=+1485.678934540" watchObservedRunningTime="2025-11-22 05:12:13.119087568 +0000 UTC m=+1485.693504095" Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.135362 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/178832e0-386a-4942-9c1f-b51959d4f455-operator-scripts\") pod \"nova-cell1-cce9-account-create-qjwmx\" (UID: \"178832e0-386a-4942-9c1f-b51959d4f455\") " pod="openstack/nova-cell1-cce9-account-create-qjwmx" Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.135425 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtj44\" (UniqueName: \"kubernetes.io/projected/178832e0-386a-4942-9c1f-b51959d4f455-kube-api-access-mtj44\") pod \"nova-cell1-cce9-account-create-qjwmx\" (UID: \"178832e0-386a-4942-9c1f-b51959d4f455\") " pod="openstack/nova-cell1-cce9-account-create-qjwmx" Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.237505 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/178832e0-386a-4942-9c1f-b51959d4f455-operator-scripts\") pod \"nova-cell1-cce9-account-create-qjwmx\" (UID: \"178832e0-386a-4942-9c1f-b51959d4f455\") " pod="openstack/nova-cell1-cce9-account-create-qjwmx" Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.237591 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtj44\" (UniqueName: \"kubernetes.io/projected/178832e0-386a-4942-9c1f-b51959d4f455-kube-api-access-mtj44\") pod \"nova-cell1-cce9-account-create-qjwmx\" (UID: \"178832e0-386a-4942-9c1f-b51959d4f455\") " pod="openstack/nova-cell1-cce9-account-create-qjwmx" Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.244878 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/178832e0-386a-4942-9c1f-b51959d4f455-operator-scripts\") pod \"nova-cell1-cce9-account-create-qjwmx\" (UID: \"178832e0-386a-4942-9c1f-b51959d4f455\") " pod="openstack/nova-cell1-cce9-account-create-qjwmx" Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.297845 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8vsln"] Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.308661 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtj44\" (UniqueName: \"kubernetes.io/projected/178832e0-386a-4942-9c1f-b51959d4f455-kube-api-access-mtj44\") pod \"nova-cell1-cce9-account-create-qjwmx\" (UID: \"178832e0-386a-4942-9c1f-b51959d4f455\") " pod="openstack/nova-cell1-cce9-account-create-qjwmx" Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.331587 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-d92ms"] Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.372781 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cce9-account-create-qjwmx" Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.636459 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-b75c-account-create-ztmdp"] Nov 22 05:12:13 crc kubenswrapper[4818]: W1122 05:12:13.652689 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e0926ba_d1e1_470f_91a0_221471d29797.slice/crio-24bafa5302e82e3d2ac6f6eb6a8656dd31dd634cc29713238f82e23876428e0d WatchSource:0}: Error finding container 24bafa5302e82e3d2ac6f6eb6a8656dd31dd634cc29713238f82e23876428e0d: Status 404 returned error can't find the container with id 24bafa5302e82e3d2ac6f6eb6a8656dd31dd634cc29713238f82e23876428e0d Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.802587 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-5njh6"] Nov 22 05:12:13 crc kubenswrapper[4818]: W1122 05:12:13.807724 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a04faa3_41d0_4e96_8957_097cba1b5154.slice/crio-4e93a596ebbcc128a7a271e26c40f8302f68ee64fb04f07adb55138fac324f24 WatchSource:0}: Error finding container 4e93a596ebbcc128a7a271e26c40f8302f68ee64fb04f07adb55138fac324f24: Status 404 returned error can't find the container with id 4e93a596ebbcc128a7a271e26c40f8302f68ee64fb04f07adb55138fac324f24 Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.855628 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-36d8-account-create-bwv78"] Nov 22 05:12:13 crc kubenswrapper[4818]: I1122 05:12:13.916761 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cce9-account-create-qjwmx"] Nov 22 05:12:13 crc kubenswrapper[4818]: W1122 05:12:13.923030 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod178832e0_386a_4942_9c1f_b51959d4f455.slice/crio-7ec77bbd7ed5520e5ddad268d6fbae56d475436aec97a714374a29b49b39a14d WatchSource:0}: Error finding container 7ec77bbd7ed5520e5ddad268d6fbae56d475436aec97a714374a29b49b39a14d: Status 404 returned error can't find the container with id 7ec77bbd7ed5520e5ddad268d6fbae56d475436aec97a714374a29b49b39a14d Nov 22 05:12:14 crc kubenswrapper[4818]: I1122 05:12:14.088072 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-36d8-account-create-bwv78" event={"ID":"f5382656-d791-455a-b7a4-925c2b37d399","Type":"ContainerStarted","Data":"d9eddc8e6decf12a4647e9e0230c7987c1f6f95be8e46a19bd2211a00ab18ba7"} Nov 22 05:12:14 crc kubenswrapper[4818]: I1122 05:12:14.091458 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8vsln" event={"ID":"39310079-a73f-4fe1-854e-aacffbddcabb","Type":"ContainerStarted","Data":"3f03b336d1afc3c9ecdd9f5afb3b508da56c440e205176a76eeb5c1be9576b09"} Nov 22 05:12:14 crc kubenswrapper[4818]: I1122 05:12:14.091517 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8vsln" event={"ID":"39310079-a73f-4fe1-854e-aacffbddcabb","Type":"ContainerStarted","Data":"8f44d8d65a8057c645e1655dfce1ab9d1b373eeacc5a1991c764d15320bbcbbe"} Nov 22 05:12:14 crc kubenswrapper[4818]: I1122 05:12:14.096851 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cce9-account-create-qjwmx" event={"ID":"178832e0-386a-4942-9c1f-b51959d4f455","Type":"ContainerStarted","Data":"7ec77bbd7ed5520e5ddad268d6fbae56d475436aec97a714374a29b49b39a14d"} Nov 22 05:12:14 crc kubenswrapper[4818]: I1122 05:12:14.098476 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b75c-account-create-ztmdp" event={"ID":"6e0926ba-d1e1-470f-91a0-221471d29797","Type":"ContainerStarted","Data":"24bafa5302e82e3d2ac6f6eb6a8656dd31dd634cc29713238f82e23876428e0d"} Nov 22 05:12:14 crc kubenswrapper[4818]: I1122 05:12:14.099989 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5njh6" event={"ID":"5a04faa3-41d0-4e96-8957-097cba1b5154","Type":"ContainerStarted","Data":"4e93a596ebbcc128a7a271e26c40f8302f68ee64fb04f07adb55138fac324f24"} Nov 22 05:12:14 crc kubenswrapper[4818]: I1122 05:12:14.101995 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-d92ms" event={"ID":"057a60f1-d0da-4590-ab6c-4de6cf4c79d1","Type":"ContainerStarted","Data":"f91a1b449e7bb6c68f7e492764683b0d1a2ab0b6e3abcd72b853dd558b3eeb8c"} Nov 22 05:12:14 crc kubenswrapper[4818]: I1122 05:12:14.102055 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-d92ms" event={"ID":"057a60f1-d0da-4590-ab6c-4de6cf4c79d1","Type":"ContainerStarted","Data":"727a607098b5bcdfb2f618dbca9059fdc94b4110b0691abbb2ecc848de3173d5"} Nov 22 05:12:14 crc kubenswrapper[4818]: I1122 05:12:14.108756 4818 generic.go:334] "Generic (PLEG): container finished" podID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" containerID="e5a594e737babf68be143838f05e73d933ece2f93df8f917e60d57e3fa0da9e9" exitCode=0 Nov 22 05:12:14 crc kubenswrapper[4818]: I1122 05:12:14.108810 4818 generic.go:334] "Generic (PLEG): container finished" podID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" containerID="adb1fc6502127770a258a3477ac90aaadb29d71775af8433af6d2f3a90f5cc31" exitCode=2 Nov 22 05:12:14 crc kubenswrapper[4818]: I1122 05:12:14.108821 4818 generic.go:334] "Generic (PLEG): container finished" podID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" containerID="3b6026ce62631ee3a487dfa65840eb6c1a3482ca6723b042a74ea5d32ab32697" exitCode=0 Nov 22 05:12:14 crc kubenswrapper[4818]: I1122 05:12:14.108846 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"38c10a66-aae4-4e0e-839d-aa86e7d619c9","Type":"ContainerDied","Data":"e5a594e737babf68be143838f05e73d933ece2f93df8f917e60d57e3fa0da9e9"} Nov 22 05:12:14 crc kubenswrapper[4818]: I1122 05:12:14.108891 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"38c10a66-aae4-4e0e-839d-aa86e7d619c9","Type":"ContainerDied","Data":"adb1fc6502127770a258a3477ac90aaadb29d71775af8433af6d2f3a90f5cc31"} Nov 22 05:12:14 crc kubenswrapper[4818]: I1122 05:12:14.108903 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"38c10a66-aae4-4e0e-839d-aa86e7d619c9","Type":"ContainerDied","Data":"3b6026ce62631ee3a487dfa65840eb6c1a3482ca6723b042a74ea5d32ab32697"} Nov 22 05:12:14 crc kubenswrapper[4818]: I1122 05:12:14.124497 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-8vsln" podStartSLOduration=2.124476588 podStartE2EDuration="2.124476588s" podCreationTimestamp="2025-11-22 05:12:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:12:14.117699419 +0000 UTC m=+1486.692115976" watchObservedRunningTime="2025-11-22 05:12:14.124476588 +0000 UTC m=+1486.698893115" Nov 22 05:12:14 crc kubenswrapper[4818]: I1122 05:12:14.134472 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-d92ms" podStartSLOduration=2.13444547 podStartE2EDuration="2.13444547s" podCreationTimestamp="2025-11-22 05:12:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:12:14.130450045 +0000 UTC m=+1486.704866582" watchObservedRunningTime="2025-11-22 05:12:14.13444547 +0000 UTC m=+1486.708861997" Nov 22 05:12:15 crc kubenswrapper[4818]: I1122 05:12:15.118240 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b75c-account-create-ztmdp" event={"ID":"6e0926ba-d1e1-470f-91a0-221471d29797","Type":"ContainerStarted","Data":"d060af874680974cd4ce7851cfd0125d9efce86e4558652834b2cec0ba64c7fa"} Nov 22 05:12:15 crc kubenswrapper[4818]: I1122 05:12:15.120102 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5njh6" event={"ID":"5a04faa3-41d0-4e96-8957-097cba1b5154","Type":"ContainerStarted","Data":"ded3d607ac66ac24edd63abaae5ebd0d99640499448085fa891e07777e9b7753"} Nov 22 05:12:15 crc kubenswrapper[4818]: I1122 05:12:15.121814 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-36d8-account-create-bwv78" event={"ID":"f5382656-d791-455a-b7a4-925c2b37d399","Type":"ContainerStarted","Data":"d4c091ded479b158d589b6c9415e9c32744493ea4418f8cc86e5a1d57e0fcb33"} Nov 22 05:12:15 crc kubenswrapper[4818]: I1122 05:12:15.123406 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cce9-account-create-qjwmx" event={"ID":"178832e0-386a-4942-9c1f-b51959d4f455","Type":"ContainerStarted","Data":"320c6f3306f348ea0711da8ae7f46f05fead8dbe60d4fc73c16afe52ead37202"} Nov 22 05:12:15 crc kubenswrapper[4818]: I1122 05:12:15.138015 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-b75c-account-create-ztmdp" podStartSLOduration=3.137998302 podStartE2EDuration="3.137998302s" podCreationTimestamp="2025-11-22 05:12:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:12:15.134544321 +0000 UTC m=+1487.708960848" watchObservedRunningTime="2025-11-22 05:12:15.137998302 +0000 UTC m=+1487.712414819" Nov 22 05:12:15 crc kubenswrapper[4818]: I1122 05:12:15.166209 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-36d8-account-create-bwv78" podStartSLOduration=3.166187935 podStartE2EDuration="3.166187935s" podCreationTimestamp="2025-11-22 05:12:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:12:15.157837515 +0000 UTC m=+1487.732254062" watchObservedRunningTime="2025-11-22 05:12:15.166187935 +0000 UTC m=+1487.740604462" Nov 22 05:12:15 crc kubenswrapper[4818]: I1122 05:12:15.177400 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cce9-account-create-qjwmx" podStartSLOduration=3.177383281 podStartE2EDuration="3.177383281s" podCreationTimestamp="2025-11-22 05:12:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:12:15.170570941 +0000 UTC m=+1487.744987468" watchObservedRunningTime="2025-11-22 05:12:15.177383281 +0000 UTC m=+1487.751799808" Nov 22 05:12:15 crc kubenswrapper[4818]: I1122 05:12:15.191541 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-5njh6" podStartSLOduration=3.191496333 podStartE2EDuration="3.191496333s" podCreationTimestamp="2025-11-22 05:12:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:12:15.183499042 +0000 UTC m=+1487.757915569" watchObservedRunningTime="2025-11-22 05:12:15.191496333 +0000 UTC m=+1487.765912870" Nov 22 05:12:15 crc kubenswrapper[4818]: I1122 05:12:15.314044 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 22 05:12:16 crc kubenswrapper[4818]: I1122 05:12:16.133380 4818 generic.go:334] "Generic (PLEG): container finished" podID="178832e0-386a-4942-9c1f-b51959d4f455" containerID="320c6f3306f348ea0711da8ae7f46f05fead8dbe60d4fc73c16afe52ead37202" exitCode=0 Nov 22 05:12:16 crc kubenswrapper[4818]: I1122 05:12:16.133485 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cce9-account-create-qjwmx" event={"ID":"178832e0-386a-4942-9c1f-b51959d4f455","Type":"ContainerDied","Data":"320c6f3306f348ea0711da8ae7f46f05fead8dbe60d4fc73c16afe52ead37202"} Nov 22 05:12:16 crc kubenswrapper[4818]: I1122 05:12:16.135175 4818 generic.go:334] "Generic (PLEG): container finished" podID="6e0926ba-d1e1-470f-91a0-221471d29797" containerID="d060af874680974cd4ce7851cfd0125d9efce86e4558652834b2cec0ba64c7fa" exitCode=0 Nov 22 05:12:16 crc kubenswrapper[4818]: I1122 05:12:16.135231 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b75c-account-create-ztmdp" event={"ID":"6e0926ba-d1e1-470f-91a0-221471d29797","Type":"ContainerDied","Data":"d060af874680974cd4ce7851cfd0125d9efce86e4558652834b2cec0ba64c7fa"} Nov 22 05:12:16 crc kubenswrapper[4818]: I1122 05:12:16.137158 4818 generic.go:334] "Generic (PLEG): container finished" podID="5a04faa3-41d0-4e96-8957-097cba1b5154" containerID="ded3d607ac66ac24edd63abaae5ebd0d99640499448085fa891e07777e9b7753" exitCode=0 Nov 22 05:12:16 crc kubenswrapper[4818]: I1122 05:12:16.137202 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5njh6" event={"ID":"5a04faa3-41d0-4e96-8957-097cba1b5154","Type":"ContainerDied","Data":"ded3d607ac66ac24edd63abaae5ebd0d99640499448085fa891e07777e9b7753"} Nov 22 05:12:16 crc kubenswrapper[4818]: I1122 05:12:16.138878 4818 generic.go:334] "Generic (PLEG): container finished" podID="057a60f1-d0da-4590-ab6c-4de6cf4c79d1" containerID="f91a1b449e7bb6c68f7e492764683b0d1a2ab0b6e3abcd72b853dd558b3eeb8c" exitCode=0 Nov 22 05:12:16 crc kubenswrapper[4818]: I1122 05:12:16.138926 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-d92ms" event={"ID":"057a60f1-d0da-4590-ab6c-4de6cf4c79d1","Type":"ContainerDied","Data":"f91a1b449e7bb6c68f7e492764683b0d1a2ab0b6e3abcd72b853dd558b3eeb8c"} Nov 22 05:12:16 crc kubenswrapper[4818]: I1122 05:12:16.140657 4818 generic.go:334] "Generic (PLEG): container finished" podID="39310079-a73f-4fe1-854e-aacffbddcabb" containerID="3f03b336d1afc3c9ecdd9f5afb3b508da56c440e205176a76eeb5c1be9576b09" exitCode=0 Nov 22 05:12:16 crc kubenswrapper[4818]: I1122 05:12:16.140716 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8vsln" event={"ID":"39310079-a73f-4fe1-854e-aacffbddcabb","Type":"ContainerDied","Data":"3f03b336d1afc3c9ecdd9f5afb3b508da56c440e205176a76eeb5c1be9576b09"} Nov 22 05:12:16 crc kubenswrapper[4818]: I1122 05:12:16.142219 4818 generic.go:334] "Generic (PLEG): container finished" podID="f5382656-d791-455a-b7a4-925c2b37d399" containerID="d4c091ded479b158d589b6c9415e9c32744493ea4418f8cc86e5a1d57e0fcb33" exitCode=0 Nov 22 05:12:16 crc kubenswrapper[4818]: I1122 05:12:16.142265 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-36d8-account-create-bwv78" event={"ID":"f5382656-d791-455a-b7a4-925c2b37d399","Type":"ContainerDied","Data":"d4c091ded479b158d589b6c9415e9c32744493ea4418f8cc86e5a1d57e0fcb33"} Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.539484 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-d92ms" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.731202 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b75c-account-create-ztmdp" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.732151 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79wmw\" (UniqueName: \"kubernetes.io/projected/057a60f1-d0da-4590-ab6c-4de6cf4c79d1-kube-api-access-79wmw\") pod \"057a60f1-d0da-4590-ab6c-4de6cf4c79d1\" (UID: \"057a60f1-d0da-4590-ab6c-4de6cf4c79d1\") " Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.732211 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/057a60f1-d0da-4590-ab6c-4de6cf4c79d1-operator-scripts\") pod \"057a60f1-d0da-4590-ab6c-4de6cf4c79d1\" (UID: \"057a60f1-d0da-4590-ab6c-4de6cf4c79d1\") " Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.732962 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cce9-account-create-qjwmx" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.733408 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/057a60f1-d0da-4590-ab6c-4de6cf4c79d1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "057a60f1-d0da-4590-ab6c-4de6cf4c79d1" (UID: "057a60f1-d0da-4590-ab6c-4de6cf4c79d1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.741538 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/057a60f1-d0da-4590-ab6c-4de6cf4c79d1-kube-api-access-79wmw" (OuterVolumeSpecName: "kube-api-access-79wmw") pod "057a60f1-d0da-4590-ab6c-4de6cf4c79d1" (UID: "057a60f1-d0da-4590-ab6c-4de6cf4c79d1"). InnerVolumeSpecName "kube-api-access-79wmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.770836 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8vsln" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.817276 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-36d8-account-create-bwv78" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.834491 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/178832e0-386a-4942-9c1f-b51959d4f455-operator-scripts\") pod \"178832e0-386a-4942-9c1f-b51959d4f455\" (UID: \"178832e0-386a-4942-9c1f-b51959d4f455\") " Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.834775 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2xgq\" (UniqueName: \"kubernetes.io/projected/6e0926ba-d1e1-470f-91a0-221471d29797-kube-api-access-l2xgq\") pod \"6e0926ba-d1e1-470f-91a0-221471d29797\" (UID: \"6e0926ba-d1e1-470f-91a0-221471d29797\") " Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.834990 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e0926ba-d1e1-470f-91a0-221471d29797-operator-scripts\") pod \"6e0926ba-d1e1-470f-91a0-221471d29797\" (UID: \"6e0926ba-d1e1-470f-91a0-221471d29797\") " Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.844873 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtj44\" (UniqueName: \"kubernetes.io/projected/178832e0-386a-4942-9c1f-b51959d4f455-kube-api-access-mtj44\") pod \"178832e0-386a-4942-9c1f-b51959d4f455\" (UID: \"178832e0-386a-4942-9c1f-b51959d4f455\") " Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.845630 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79wmw\" (UniqueName: \"kubernetes.io/projected/057a60f1-d0da-4590-ab6c-4de6cf4c79d1-kube-api-access-79wmw\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.845651 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/057a60f1-d0da-4590-ab6c-4de6cf4c79d1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.850038 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e0926ba-d1e1-470f-91a0-221471d29797-kube-api-access-l2xgq" (OuterVolumeSpecName: "kube-api-access-l2xgq") pod "6e0926ba-d1e1-470f-91a0-221471d29797" (UID: "6e0926ba-d1e1-470f-91a0-221471d29797"). InnerVolumeSpecName "kube-api-access-l2xgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.850314 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e0926ba-d1e1-470f-91a0-221471d29797-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6e0926ba-d1e1-470f-91a0-221471d29797" (UID: "6e0926ba-d1e1-470f-91a0-221471d29797"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.850385 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/178832e0-386a-4942-9c1f-b51959d4f455-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "178832e0-386a-4942-9c1f-b51959d4f455" (UID: "178832e0-386a-4942-9c1f-b51959d4f455"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.850640 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5njh6" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.854951 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/178832e0-386a-4942-9c1f-b51959d4f455-kube-api-access-mtj44" (OuterVolumeSpecName: "kube-api-access-mtj44") pod "178832e0-386a-4942-9c1f-b51959d4f455" (UID: "178832e0-386a-4942-9c1f-b51959d4f455"). InnerVolumeSpecName "kube-api-access-mtj44". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.946542 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjnqd\" (UniqueName: \"kubernetes.io/projected/5a04faa3-41d0-4e96-8957-097cba1b5154-kube-api-access-bjnqd\") pod \"5a04faa3-41d0-4e96-8957-097cba1b5154\" (UID: \"5a04faa3-41d0-4e96-8957-097cba1b5154\") " Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.946599 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zqcm\" (UniqueName: \"kubernetes.io/projected/39310079-a73f-4fe1-854e-aacffbddcabb-kube-api-access-4zqcm\") pod \"39310079-a73f-4fe1-854e-aacffbddcabb\" (UID: \"39310079-a73f-4fe1-854e-aacffbddcabb\") " Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.946647 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9jbx\" (UniqueName: \"kubernetes.io/projected/f5382656-d791-455a-b7a4-925c2b37d399-kube-api-access-f9jbx\") pod \"f5382656-d791-455a-b7a4-925c2b37d399\" (UID: \"f5382656-d791-455a-b7a4-925c2b37d399\") " Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.946776 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a04faa3-41d0-4e96-8957-097cba1b5154-operator-scripts\") pod \"5a04faa3-41d0-4e96-8957-097cba1b5154\" (UID: \"5a04faa3-41d0-4e96-8957-097cba1b5154\") " Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.946859 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/39310079-a73f-4fe1-854e-aacffbddcabb-operator-scripts\") pod \"39310079-a73f-4fe1-854e-aacffbddcabb\" (UID: \"39310079-a73f-4fe1-854e-aacffbddcabb\") " Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.946900 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5382656-d791-455a-b7a4-925c2b37d399-operator-scripts\") pod \"f5382656-d791-455a-b7a4-925c2b37d399\" (UID: \"f5382656-d791-455a-b7a4-925c2b37d399\") " Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.947418 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a04faa3-41d0-4e96-8957-097cba1b5154-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5a04faa3-41d0-4e96-8957-097cba1b5154" (UID: "5a04faa3-41d0-4e96-8957-097cba1b5154"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.947433 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39310079-a73f-4fe1-854e-aacffbddcabb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "39310079-a73f-4fe1-854e-aacffbddcabb" (UID: "39310079-a73f-4fe1-854e-aacffbddcabb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.947475 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5382656-d791-455a-b7a4-925c2b37d399-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f5382656-d791-455a-b7a4-925c2b37d399" (UID: "f5382656-d791-455a-b7a4-925c2b37d399"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.947791 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/178832e0-386a-4942-9c1f-b51959d4f455-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.947818 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2xgq\" (UniqueName: \"kubernetes.io/projected/6e0926ba-d1e1-470f-91a0-221471d29797-kube-api-access-l2xgq\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.947834 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e0926ba-d1e1-470f-91a0-221471d29797-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.947848 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/39310079-a73f-4fe1-854e-aacffbddcabb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.947859 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5382656-d791-455a-b7a4-925c2b37d399-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.947871 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtj44\" (UniqueName: \"kubernetes.io/projected/178832e0-386a-4942-9c1f-b51959d4f455-kube-api-access-mtj44\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.947883 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a04faa3-41d0-4e96-8957-097cba1b5154-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.950228 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5382656-d791-455a-b7a4-925c2b37d399-kube-api-access-f9jbx" (OuterVolumeSpecName: "kube-api-access-f9jbx") pod "f5382656-d791-455a-b7a4-925c2b37d399" (UID: "f5382656-d791-455a-b7a4-925c2b37d399"). InnerVolumeSpecName "kube-api-access-f9jbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.950762 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39310079-a73f-4fe1-854e-aacffbddcabb-kube-api-access-4zqcm" (OuterVolumeSpecName: "kube-api-access-4zqcm") pod "39310079-a73f-4fe1-854e-aacffbddcabb" (UID: "39310079-a73f-4fe1-854e-aacffbddcabb"). InnerVolumeSpecName "kube-api-access-4zqcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:12:17 crc kubenswrapper[4818]: I1122 05:12:17.952039 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a04faa3-41d0-4e96-8957-097cba1b5154-kube-api-access-bjnqd" (OuterVolumeSpecName: "kube-api-access-bjnqd") pod "5a04faa3-41d0-4e96-8957-097cba1b5154" (UID: "5a04faa3-41d0-4e96-8957-097cba1b5154"). InnerVolumeSpecName "kube-api-access-bjnqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.051506 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjnqd\" (UniqueName: \"kubernetes.io/projected/5a04faa3-41d0-4e96-8957-097cba1b5154-kube-api-access-bjnqd\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.051541 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zqcm\" (UniqueName: \"kubernetes.io/projected/39310079-a73f-4fe1-854e-aacffbddcabb-kube-api-access-4zqcm\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.051550 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9jbx\" (UniqueName: \"kubernetes.io/projected/f5382656-d791-455a-b7a4-925c2b37d399-kube-api-access-f9jbx\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.160159 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-d92ms" event={"ID":"057a60f1-d0da-4590-ab6c-4de6cf4c79d1","Type":"ContainerDied","Data":"727a607098b5bcdfb2f618dbca9059fdc94b4110b0691abbb2ecc848de3173d5"} Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.160480 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="727a607098b5bcdfb2f618dbca9059fdc94b4110b0691abbb2ecc848de3173d5" Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.160170 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-d92ms" Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.162504 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8vsln" event={"ID":"39310079-a73f-4fe1-854e-aacffbddcabb","Type":"ContainerDied","Data":"8f44d8d65a8057c645e1655dfce1ab9d1b373eeacc5a1991c764d15320bbcbbe"} Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.162665 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f44d8d65a8057c645e1655dfce1ab9d1b373eeacc5a1991c764d15320bbcbbe" Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.162560 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8vsln" Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.164394 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-36d8-account-create-bwv78" Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.164399 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-36d8-account-create-bwv78" event={"ID":"f5382656-d791-455a-b7a4-925c2b37d399","Type":"ContainerDied","Data":"d9eddc8e6decf12a4647e9e0230c7987c1f6f95be8e46a19bd2211a00ab18ba7"} Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.164445 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9eddc8e6decf12a4647e9e0230c7987c1f6f95be8e46a19bd2211a00ab18ba7" Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.165835 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cce9-account-create-qjwmx" event={"ID":"178832e0-386a-4942-9c1f-b51959d4f455","Type":"ContainerDied","Data":"7ec77bbd7ed5520e5ddad268d6fbae56d475436aec97a714374a29b49b39a14d"} Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.165864 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ec77bbd7ed5520e5ddad268d6fbae56d475436aec97a714374a29b49b39a14d" Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.165843 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cce9-account-create-qjwmx" Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.167593 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b75c-account-create-ztmdp" event={"ID":"6e0926ba-d1e1-470f-91a0-221471d29797","Type":"ContainerDied","Data":"24bafa5302e82e3d2ac6f6eb6a8656dd31dd634cc29713238f82e23876428e0d"} Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.167631 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24bafa5302e82e3d2ac6f6eb6a8656dd31dd634cc29713238f82e23876428e0d" Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.167614 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b75c-account-create-ztmdp" Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.168780 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5njh6" event={"ID":"5a04faa3-41d0-4e96-8957-097cba1b5154","Type":"ContainerDied","Data":"4e93a596ebbcc128a7a271e26c40f8302f68ee64fb04f07adb55138fac324f24"} Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.168813 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e93a596ebbcc128a7a271e26c40f8302f68ee64fb04f07adb55138fac324f24" Nov 22 05:12:18 crc kubenswrapper[4818]: I1122 05:12:18.168848 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5njh6" Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.180582 4818 generic.go:334] "Generic (PLEG): container finished" podID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" containerID="e0c6e0e1acc116e5e9152826b2ef6dff3b73a7d824ed2fa9dba73782f1d203d6" exitCode=0 Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.180942 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"38c10a66-aae4-4e0e-839d-aa86e7d619c9","Type":"ContainerDied","Data":"e0c6e0e1acc116e5e9152826b2ef6dff3b73a7d824ed2fa9dba73782f1d203d6"} Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.442754 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.576660 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-scripts\") pod \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.576729 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-ceilometer-tls-certs\") pod \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.576824 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/38c10a66-aae4-4e0e-839d-aa86e7d619c9-run-httpd\") pod \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.576869 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/38c10a66-aae4-4e0e-839d-aa86e7d619c9-log-httpd\") pod \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.576937 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-combined-ca-bundle\") pod \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.577284 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38c10a66-aae4-4e0e-839d-aa86e7d619c9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "38c10a66-aae4-4e0e-839d-aa86e7d619c9" (UID: "38c10a66-aae4-4e0e-839d-aa86e7d619c9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.577029 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2jz7\" (UniqueName: \"kubernetes.io/projected/38c10a66-aae4-4e0e-839d-aa86e7d619c9-kube-api-access-j2jz7\") pod \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.577599 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-sg-core-conf-yaml\") pod \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.577472 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38c10a66-aae4-4e0e-839d-aa86e7d619c9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "38c10a66-aae4-4e0e-839d-aa86e7d619c9" (UID: "38c10a66-aae4-4e0e-839d-aa86e7d619c9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.577640 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-config-data\") pod \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\" (UID: \"38c10a66-aae4-4e0e-839d-aa86e7d619c9\") " Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.578548 4818 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/38c10a66-aae4-4e0e-839d-aa86e7d619c9-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.578704 4818 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/38c10a66-aae4-4e0e-839d-aa86e7d619c9-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.608132 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-scripts" (OuterVolumeSpecName: "scripts") pod "38c10a66-aae4-4e0e-839d-aa86e7d619c9" (UID: "38c10a66-aae4-4e0e-839d-aa86e7d619c9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.609484 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38c10a66-aae4-4e0e-839d-aa86e7d619c9-kube-api-access-j2jz7" (OuterVolumeSpecName: "kube-api-access-j2jz7") pod "38c10a66-aae4-4e0e-839d-aa86e7d619c9" (UID: "38c10a66-aae4-4e0e-839d-aa86e7d619c9"). InnerVolumeSpecName "kube-api-access-j2jz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.612031 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "38c10a66-aae4-4e0e-839d-aa86e7d619c9" (UID: "38c10a66-aae4-4e0e-839d-aa86e7d619c9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.643421 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "38c10a66-aae4-4e0e-839d-aa86e7d619c9" (UID: "38c10a66-aae4-4e0e-839d-aa86e7d619c9"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.668377 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "38c10a66-aae4-4e0e-839d-aa86e7d619c9" (UID: "38c10a66-aae4-4e0e-839d-aa86e7d619c9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.680174 4818 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.680214 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.680229 4818 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.680241 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.680270 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2jz7\" (UniqueName: \"kubernetes.io/projected/38c10a66-aae4-4e0e-839d-aa86e7d619c9-kube-api-access-j2jz7\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.697998 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-config-data" (OuterVolumeSpecName: "config-data") pod "38c10a66-aae4-4e0e-839d-aa86e7d619c9" (UID: "38c10a66-aae4-4e0e-839d-aa86e7d619c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:19 crc kubenswrapper[4818]: I1122 05:12:19.781394 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38c10a66-aae4-4e0e-839d-aa86e7d619c9-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.191453 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"38c10a66-aae4-4e0e-839d-aa86e7d619c9","Type":"ContainerDied","Data":"2768445f3c85d14c8f7211723d5aec5a43eca4bbbe27933089a82098b9142edf"} Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.191528 4818 scope.go:117] "RemoveContainer" containerID="e5a594e737babf68be143838f05e73d933ece2f93df8f917e60d57e3fa0da9e9" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.191517 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.210121 4818 scope.go:117] "RemoveContainer" containerID="adb1fc6502127770a258a3477ac90aaadb29d71775af8433af6d2f3a90f5cc31" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.230902 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.231694 4818 scope.go:117] "RemoveContainer" containerID="3b6026ce62631ee3a487dfa65840eb6c1a3482ca6723b042a74ea5d32ab32697" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.239456 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.247687 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:20 crc kubenswrapper[4818]: E1122 05:12:20.248032 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="057a60f1-d0da-4590-ab6c-4de6cf4c79d1" containerName="mariadb-database-create" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.248044 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="057a60f1-d0da-4590-ab6c-4de6cf4c79d1" containerName="mariadb-database-create" Nov 22 05:12:20 crc kubenswrapper[4818]: E1122 05:12:20.248057 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5382656-d791-455a-b7a4-925c2b37d399" containerName="mariadb-account-create" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.248063 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5382656-d791-455a-b7a4-925c2b37d399" containerName="mariadb-account-create" Nov 22 05:12:20 crc kubenswrapper[4818]: E1122 05:12:20.248085 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a04faa3-41d0-4e96-8957-097cba1b5154" containerName="mariadb-database-create" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.248091 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a04faa3-41d0-4e96-8957-097cba1b5154" containerName="mariadb-database-create" Nov 22 05:12:20 crc kubenswrapper[4818]: E1122 05:12:20.248102 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e0926ba-d1e1-470f-91a0-221471d29797" containerName="mariadb-account-create" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.248108 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e0926ba-d1e1-470f-91a0-221471d29797" containerName="mariadb-account-create" Nov 22 05:12:20 crc kubenswrapper[4818]: E1122 05:12:20.248115 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" containerName="ceilometer-notification-agent" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.248123 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" containerName="ceilometer-notification-agent" Nov 22 05:12:20 crc kubenswrapper[4818]: E1122 05:12:20.248133 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" containerName="sg-core" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.248139 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" containerName="sg-core" Nov 22 05:12:20 crc kubenswrapper[4818]: E1122 05:12:20.248159 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" containerName="proxy-httpd" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.248165 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" containerName="proxy-httpd" Nov 22 05:12:20 crc kubenswrapper[4818]: E1122 05:12:20.248173 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39310079-a73f-4fe1-854e-aacffbddcabb" containerName="mariadb-database-create" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.248179 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="39310079-a73f-4fe1-854e-aacffbddcabb" containerName="mariadb-database-create" Nov 22 05:12:20 crc kubenswrapper[4818]: E1122 05:12:20.248186 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" containerName="ceilometer-central-agent" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.248191 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" containerName="ceilometer-central-agent" Nov 22 05:12:20 crc kubenswrapper[4818]: E1122 05:12:20.248202 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="178832e0-386a-4942-9c1f-b51959d4f455" containerName="mariadb-account-create" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.248208 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="178832e0-386a-4942-9c1f-b51959d4f455" containerName="mariadb-account-create" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.248539 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e0926ba-d1e1-470f-91a0-221471d29797" containerName="mariadb-account-create" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.248552 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5382656-d791-455a-b7a4-925c2b37d399" containerName="mariadb-account-create" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.248561 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="178832e0-386a-4942-9c1f-b51959d4f455" containerName="mariadb-account-create" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.248569 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a04faa3-41d0-4e96-8957-097cba1b5154" containerName="mariadb-database-create" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.248581 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" containerName="ceilometer-central-agent" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.248588 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="39310079-a73f-4fe1-854e-aacffbddcabb" containerName="mariadb-database-create" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.248595 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="057a60f1-d0da-4590-ab6c-4de6cf4c79d1" containerName="mariadb-database-create" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.248605 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" containerName="ceilometer-notification-agent" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.248615 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" containerName="sg-core" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.248623 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" containerName="proxy-httpd" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.250138 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.258412 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.259197 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.259531 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.260223 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.262505 4818 scope.go:117] "RemoveContainer" containerID="e0c6e0e1acc116e5e9152826b2ef6dff3b73a7d824ed2fa9dba73782f1d203d6" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.318123 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38c10a66-aae4-4e0e-839d-aa86e7d619c9" path="/var/lib/kubelet/pods/38c10a66-aae4-4e0e-839d-aa86e7d619c9/volumes" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.387363 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:20 crc kubenswrapper[4818]: E1122 05:12:20.387934 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceilometer-tls-certs combined-ca-bundle config-data kube-api-access-8ptp6 log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="f879d527-8a60-4f17-a77f-3ea486f0710a" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.400386 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-config-data\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.400477 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f879d527-8a60-4f17-a77f-3ea486f0710a-run-httpd\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.400520 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.400556 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f879d527-8a60-4f17-a77f-3ea486f0710a-log-httpd\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.400622 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.400654 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-scripts\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.400674 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ptp6\" (UniqueName: \"kubernetes.io/projected/f879d527-8a60-4f17-a77f-3ea486f0710a-kube-api-access-8ptp6\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.400705 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.502151 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-config-data\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.502232 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f879d527-8a60-4f17-a77f-3ea486f0710a-run-httpd\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.502306 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.502328 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f879d527-8a60-4f17-a77f-3ea486f0710a-log-httpd\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.502828 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f879d527-8a60-4f17-a77f-3ea486f0710a-run-httpd\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.502874 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f879d527-8a60-4f17-a77f-3ea486f0710a-log-httpd\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.503081 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.503116 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-scripts\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.503131 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ptp6\" (UniqueName: \"kubernetes.io/projected/f879d527-8a60-4f17-a77f-3ea486f0710a-kube-api-access-8ptp6\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.503158 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.513878 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-config-data\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.514838 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-scripts\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.514993 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.515951 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.518423 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:20 crc kubenswrapper[4818]: I1122 05:12:20.525408 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ptp6\" (UniqueName: \"kubernetes.io/projected/f879d527-8a60-4f17-a77f-3ea486f0710a-kube-api-access-8ptp6\") pod \"ceilometer-0\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " pod="openstack/ceilometer-0" Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.200219 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.210206 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.314961 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f879d527-8a60-4f17-a77f-3ea486f0710a-run-httpd\") pod \"f879d527-8a60-4f17-a77f-3ea486f0710a\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.315047 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-ceilometer-tls-certs\") pod \"f879d527-8a60-4f17-a77f-3ea486f0710a\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.315086 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-combined-ca-bundle\") pod \"f879d527-8a60-4f17-a77f-3ea486f0710a\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.315119 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ptp6\" (UniqueName: \"kubernetes.io/projected/f879d527-8a60-4f17-a77f-3ea486f0710a-kube-api-access-8ptp6\") pod \"f879d527-8a60-4f17-a77f-3ea486f0710a\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.315205 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-config-data\") pod \"f879d527-8a60-4f17-a77f-3ea486f0710a\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.315267 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f879d527-8a60-4f17-a77f-3ea486f0710a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f879d527-8a60-4f17-a77f-3ea486f0710a" (UID: "f879d527-8a60-4f17-a77f-3ea486f0710a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.315278 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-scripts\") pod \"f879d527-8a60-4f17-a77f-3ea486f0710a\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.315325 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f879d527-8a60-4f17-a77f-3ea486f0710a-log-httpd\") pod \"f879d527-8a60-4f17-a77f-3ea486f0710a\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.315346 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-sg-core-conf-yaml\") pod \"f879d527-8a60-4f17-a77f-3ea486f0710a\" (UID: \"f879d527-8a60-4f17-a77f-3ea486f0710a\") " Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.315775 4818 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f879d527-8a60-4f17-a77f-3ea486f0710a-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.316571 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f879d527-8a60-4f17-a77f-3ea486f0710a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f879d527-8a60-4f17-a77f-3ea486f0710a" (UID: "f879d527-8a60-4f17-a77f-3ea486f0710a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.321178 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-scripts" (OuterVolumeSpecName: "scripts") pod "f879d527-8a60-4f17-a77f-3ea486f0710a" (UID: "f879d527-8a60-4f17-a77f-3ea486f0710a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.321561 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "f879d527-8a60-4f17-a77f-3ea486f0710a" (UID: "f879d527-8a60-4f17-a77f-3ea486f0710a"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.324324 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-config-data" (OuterVolumeSpecName: "config-data") pod "f879d527-8a60-4f17-a77f-3ea486f0710a" (UID: "f879d527-8a60-4f17-a77f-3ea486f0710a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.324455 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f879d527-8a60-4f17-a77f-3ea486f0710a-kube-api-access-8ptp6" (OuterVolumeSpecName: "kube-api-access-8ptp6") pod "f879d527-8a60-4f17-a77f-3ea486f0710a" (UID: "f879d527-8a60-4f17-a77f-3ea486f0710a"). InnerVolumeSpecName "kube-api-access-8ptp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.332561 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f879d527-8a60-4f17-a77f-3ea486f0710a" (UID: "f879d527-8a60-4f17-a77f-3ea486f0710a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.332595 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f879d527-8a60-4f17-a77f-3ea486f0710a" (UID: "f879d527-8a60-4f17-a77f-3ea486f0710a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.417387 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.417416 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.417425 4818 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f879d527-8a60-4f17-a77f-3ea486f0710a-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.417434 4818 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.417444 4818 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.417453 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f879d527-8a60-4f17-a77f-3ea486f0710a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:21 crc kubenswrapper[4818]: I1122 05:12:21.417463 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ptp6\" (UniqueName: \"kubernetes.io/projected/f879d527-8a60-4f17-a77f-3ea486f0710a-kube-api-access-8ptp6\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.209763 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.270986 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.287129 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.303787 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f879d527-8a60-4f17-a77f-3ea486f0710a" path="/var/lib/kubelet/pods/f879d527-8a60-4f17-a77f-3ea486f0710a/volumes" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.305812 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.308343 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.309998 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.312478 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.313316 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.315502 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.433794 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.433919 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbdzq\" (UniqueName: \"kubernetes.io/projected/501083a5-d8f1-4171-8eca-c331c883ddb9-kube-api-access-gbdzq\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.433943 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.433971 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-scripts\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.433989 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/501083a5-d8f1-4171-8eca-c331c883ddb9-run-httpd\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.434292 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/501083a5-d8f1-4171-8eca-c331c883ddb9-log-httpd\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.434349 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-config-data\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.434365 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.536123 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbdzq\" (UniqueName: \"kubernetes.io/projected/501083a5-d8f1-4171-8eca-c331c883ddb9-kube-api-access-gbdzq\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.536397 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.536434 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-scripts\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.536451 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/501083a5-d8f1-4171-8eca-c331c883ddb9-run-httpd\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.536488 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/501083a5-d8f1-4171-8eca-c331c883ddb9-log-httpd\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.536505 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-config-data\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.536525 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.536577 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.537121 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/501083a5-d8f1-4171-8eca-c331c883ddb9-log-httpd\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.537523 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/501083a5-d8f1-4171-8eca-c331c883ddb9-run-httpd\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.541899 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.542433 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.546223 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.547310 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-scripts\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.547302 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-config-data\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.555448 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbdzq\" (UniqueName: \"kubernetes.io/projected/501083a5-d8f1-4171-8eca-c331c883ddb9-kube-api-access-gbdzq\") pod \"ceilometer-0\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.624228 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.930929 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.990188 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xwlll"] Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.991471 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xwlll" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.997109 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.997272 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-rd7lq" Nov 22 05:12:22 crc kubenswrapper[4818]: I1122 05:12:22.997322 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 22 05:12:23 crc kubenswrapper[4818]: I1122 05:12:23.013676 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xwlll"] Nov 22 05:12:23 crc kubenswrapper[4818]: I1122 05:12:23.146973 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fbaf9a-fd45-4f17-a114-b64695d2456d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xwlll\" (UID: \"00fbaf9a-fd45-4f17-a114-b64695d2456d\") " pod="openstack/nova-cell0-conductor-db-sync-xwlll" Nov 22 05:12:23 crc kubenswrapper[4818]: I1122 05:12:23.147064 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7pb8\" (UniqueName: \"kubernetes.io/projected/00fbaf9a-fd45-4f17-a114-b64695d2456d-kube-api-access-p7pb8\") pod \"nova-cell0-conductor-db-sync-xwlll\" (UID: \"00fbaf9a-fd45-4f17-a114-b64695d2456d\") " pod="openstack/nova-cell0-conductor-db-sync-xwlll" Nov 22 05:12:23 crc kubenswrapper[4818]: I1122 05:12:23.147445 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00fbaf9a-fd45-4f17-a114-b64695d2456d-scripts\") pod \"nova-cell0-conductor-db-sync-xwlll\" (UID: \"00fbaf9a-fd45-4f17-a114-b64695d2456d\") " pod="openstack/nova-cell0-conductor-db-sync-xwlll" Nov 22 05:12:23 crc kubenswrapper[4818]: I1122 05:12:23.147487 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00fbaf9a-fd45-4f17-a114-b64695d2456d-config-data\") pod \"nova-cell0-conductor-db-sync-xwlll\" (UID: \"00fbaf9a-fd45-4f17-a114-b64695d2456d\") " pod="openstack/nova-cell0-conductor-db-sync-xwlll" Nov 22 05:12:23 crc kubenswrapper[4818]: I1122 05:12:23.219192 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"501083a5-d8f1-4171-8eca-c331c883ddb9","Type":"ContainerStarted","Data":"64816fd7802afcfb11caa78f0af1da023c2d1214034f77608782262c03620d55"} Nov 22 05:12:23 crc kubenswrapper[4818]: I1122 05:12:23.249171 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fbaf9a-fd45-4f17-a114-b64695d2456d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xwlll\" (UID: \"00fbaf9a-fd45-4f17-a114-b64695d2456d\") " pod="openstack/nova-cell0-conductor-db-sync-xwlll" Nov 22 05:12:23 crc kubenswrapper[4818]: I1122 05:12:23.249246 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7pb8\" (UniqueName: \"kubernetes.io/projected/00fbaf9a-fd45-4f17-a114-b64695d2456d-kube-api-access-p7pb8\") pod \"nova-cell0-conductor-db-sync-xwlll\" (UID: \"00fbaf9a-fd45-4f17-a114-b64695d2456d\") " pod="openstack/nova-cell0-conductor-db-sync-xwlll" Nov 22 05:12:23 crc kubenswrapper[4818]: I1122 05:12:23.249313 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00fbaf9a-fd45-4f17-a114-b64695d2456d-scripts\") pod \"nova-cell0-conductor-db-sync-xwlll\" (UID: \"00fbaf9a-fd45-4f17-a114-b64695d2456d\") " pod="openstack/nova-cell0-conductor-db-sync-xwlll" Nov 22 05:12:23 crc kubenswrapper[4818]: I1122 05:12:23.249330 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00fbaf9a-fd45-4f17-a114-b64695d2456d-config-data\") pod \"nova-cell0-conductor-db-sync-xwlll\" (UID: \"00fbaf9a-fd45-4f17-a114-b64695d2456d\") " pod="openstack/nova-cell0-conductor-db-sync-xwlll" Nov 22 05:12:23 crc kubenswrapper[4818]: I1122 05:12:23.253844 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fbaf9a-fd45-4f17-a114-b64695d2456d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xwlll\" (UID: \"00fbaf9a-fd45-4f17-a114-b64695d2456d\") " pod="openstack/nova-cell0-conductor-db-sync-xwlll" Nov 22 05:12:23 crc kubenswrapper[4818]: I1122 05:12:23.254366 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00fbaf9a-fd45-4f17-a114-b64695d2456d-scripts\") pod \"nova-cell0-conductor-db-sync-xwlll\" (UID: \"00fbaf9a-fd45-4f17-a114-b64695d2456d\") " pod="openstack/nova-cell0-conductor-db-sync-xwlll" Nov 22 05:12:23 crc kubenswrapper[4818]: I1122 05:12:23.254449 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00fbaf9a-fd45-4f17-a114-b64695d2456d-config-data\") pod \"nova-cell0-conductor-db-sync-xwlll\" (UID: \"00fbaf9a-fd45-4f17-a114-b64695d2456d\") " pod="openstack/nova-cell0-conductor-db-sync-xwlll" Nov 22 05:12:23 crc kubenswrapper[4818]: I1122 05:12:23.265678 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7pb8\" (UniqueName: \"kubernetes.io/projected/00fbaf9a-fd45-4f17-a114-b64695d2456d-kube-api-access-p7pb8\") pod \"nova-cell0-conductor-db-sync-xwlll\" (UID: \"00fbaf9a-fd45-4f17-a114-b64695d2456d\") " pod="openstack/nova-cell0-conductor-db-sync-xwlll" Nov 22 05:12:23 crc kubenswrapper[4818]: I1122 05:12:23.311450 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xwlll" Nov 22 05:12:23 crc kubenswrapper[4818]: I1122 05:12:23.849278 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xwlll"] Nov 22 05:12:24 crc kubenswrapper[4818]: I1122 05:12:24.229635 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xwlll" event={"ID":"00fbaf9a-fd45-4f17-a114-b64695d2456d","Type":"ContainerStarted","Data":"327a32e347b746540aba0a44a89bf1e9e6d45d6a5a806bf8c70da06052a8389b"} Nov 22 05:12:25 crc kubenswrapper[4818]: I1122 05:12:25.242698 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"501083a5-d8f1-4171-8eca-c331c883ddb9","Type":"ContainerStarted","Data":"35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c"} Nov 22 05:12:26 crc kubenswrapper[4818]: I1122 05:12:26.261583 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"501083a5-d8f1-4171-8eca-c331c883ddb9","Type":"ContainerStarted","Data":"a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6"} Nov 22 05:12:27 crc kubenswrapper[4818]: I1122 05:12:27.271392 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"501083a5-d8f1-4171-8eca-c331c883ddb9","Type":"ContainerStarted","Data":"3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2"} Nov 22 05:12:32 crc kubenswrapper[4818]: I1122 05:12:32.331592 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"501083a5-d8f1-4171-8eca-c331c883ddb9","Type":"ContainerStarted","Data":"8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344"} Nov 22 05:12:32 crc kubenswrapper[4818]: I1122 05:12:32.335242 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 05:12:32 crc kubenswrapper[4818]: I1122 05:12:32.338831 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xwlll" event={"ID":"00fbaf9a-fd45-4f17-a114-b64695d2456d","Type":"ContainerStarted","Data":"cabefd1cca18ebbda0e78b29b6efb1274346a30db9cdbd8a9d7dca2b2bb35068"} Nov 22 05:12:32 crc kubenswrapper[4818]: I1122 05:12:32.369914 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.254100976 podStartE2EDuration="10.369888432s" podCreationTimestamp="2025-11-22 05:12:22 +0000 UTC" firstStartedPulling="2025-11-22 05:12:22.941527255 +0000 UTC m=+1495.515943782" lastFinishedPulling="2025-11-22 05:12:31.057314711 +0000 UTC m=+1503.631731238" observedRunningTime="2025-11-22 05:12:32.366918313 +0000 UTC m=+1504.941334890" watchObservedRunningTime="2025-11-22 05:12:32.369888432 +0000 UTC m=+1504.944304999" Nov 22 05:12:32 crc kubenswrapper[4818]: I1122 05:12:32.386533 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-xwlll" podStartSLOduration=3.189674394 podStartE2EDuration="10.38650671s" podCreationTimestamp="2025-11-22 05:12:22 +0000 UTC" firstStartedPulling="2025-11-22 05:12:23.863070434 +0000 UTC m=+1496.437486961" lastFinishedPulling="2025-11-22 05:12:31.05990275 +0000 UTC m=+1503.634319277" observedRunningTime="2025-11-22 05:12:32.385162724 +0000 UTC m=+1504.959579301" watchObservedRunningTime="2025-11-22 05:12:32.38650671 +0000 UTC m=+1504.960923247" Nov 22 05:12:43 crc kubenswrapper[4818]: I1122 05:12:43.447583 4818 generic.go:334] "Generic (PLEG): container finished" podID="00fbaf9a-fd45-4f17-a114-b64695d2456d" containerID="cabefd1cca18ebbda0e78b29b6efb1274346a30db9cdbd8a9d7dca2b2bb35068" exitCode=0 Nov 22 05:12:43 crc kubenswrapper[4818]: I1122 05:12:43.447675 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xwlll" event={"ID":"00fbaf9a-fd45-4f17-a114-b64695d2456d","Type":"ContainerDied","Data":"cabefd1cca18ebbda0e78b29b6efb1274346a30db9cdbd8a9d7dca2b2bb35068"} Nov 22 05:12:44 crc kubenswrapper[4818]: I1122 05:12:44.833702 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xwlll" Nov 22 05:12:44 crc kubenswrapper[4818]: I1122 05:12:44.993922 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7pb8\" (UniqueName: \"kubernetes.io/projected/00fbaf9a-fd45-4f17-a114-b64695d2456d-kube-api-access-p7pb8\") pod \"00fbaf9a-fd45-4f17-a114-b64695d2456d\" (UID: \"00fbaf9a-fd45-4f17-a114-b64695d2456d\") " Nov 22 05:12:44 crc kubenswrapper[4818]: I1122 05:12:44.994108 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00fbaf9a-fd45-4f17-a114-b64695d2456d-scripts\") pod \"00fbaf9a-fd45-4f17-a114-b64695d2456d\" (UID: \"00fbaf9a-fd45-4f17-a114-b64695d2456d\") " Nov 22 05:12:44 crc kubenswrapper[4818]: I1122 05:12:44.994358 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fbaf9a-fd45-4f17-a114-b64695d2456d-combined-ca-bundle\") pod \"00fbaf9a-fd45-4f17-a114-b64695d2456d\" (UID: \"00fbaf9a-fd45-4f17-a114-b64695d2456d\") " Nov 22 05:12:44 crc kubenswrapper[4818]: I1122 05:12:44.994541 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00fbaf9a-fd45-4f17-a114-b64695d2456d-config-data\") pod \"00fbaf9a-fd45-4f17-a114-b64695d2456d\" (UID: \"00fbaf9a-fd45-4f17-a114-b64695d2456d\") " Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.001630 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00fbaf9a-fd45-4f17-a114-b64695d2456d-scripts" (OuterVolumeSpecName: "scripts") pod "00fbaf9a-fd45-4f17-a114-b64695d2456d" (UID: "00fbaf9a-fd45-4f17-a114-b64695d2456d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.002146 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00fbaf9a-fd45-4f17-a114-b64695d2456d-kube-api-access-p7pb8" (OuterVolumeSpecName: "kube-api-access-p7pb8") pod "00fbaf9a-fd45-4f17-a114-b64695d2456d" (UID: "00fbaf9a-fd45-4f17-a114-b64695d2456d"). InnerVolumeSpecName "kube-api-access-p7pb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.046010 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00fbaf9a-fd45-4f17-a114-b64695d2456d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "00fbaf9a-fd45-4f17-a114-b64695d2456d" (UID: "00fbaf9a-fd45-4f17-a114-b64695d2456d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.049114 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00fbaf9a-fd45-4f17-a114-b64695d2456d-config-data" (OuterVolumeSpecName: "config-data") pod "00fbaf9a-fd45-4f17-a114-b64695d2456d" (UID: "00fbaf9a-fd45-4f17-a114-b64695d2456d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.096886 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fbaf9a-fd45-4f17-a114-b64695d2456d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.096930 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00fbaf9a-fd45-4f17-a114-b64695d2456d-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.096944 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7pb8\" (UniqueName: \"kubernetes.io/projected/00fbaf9a-fd45-4f17-a114-b64695d2456d-kube-api-access-p7pb8\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.096957 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00fbaf9a-fd45-4f17-a114-b64695d2456d-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.473556 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xwlll" event={"ID":"00fbaf9a-fd45-4f17-a114-b64695d2456d","Type":"ContainerDied","Data":"327a32e347b746540aba0a44a89bf1e9e6d45d6a5a806bf8c70da06052a8389b"} Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.473772 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="327a32e347b746540aba0a44a89bf1e9e6d45d6a5a806bf8c70da06052a8389b" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.473681 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xwlll" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.627755 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 22 05:12:45 crc kubenswrapper[4818]: E1122 05:12:45.628167 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00fbaf9a-fd45-4f17-a114-b64695d2456d" containerName="nova-cell0-conductor-db-sync" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.628353 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="00fbaf9a-fd45-4f17-a114-b64695d2456d" containerName="nova-cell0-conductor-db-sync" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.628669 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="00fbaf9a-fd45-4f17-a114-b64695d2456d" containerName="nova-cell0-conductor-db-sync" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.629365 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.631443 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.632469 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-rd7lq" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.636026 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.708245 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqms6\" (UniqueName: \"kubernetes.io/projected/5f15bdce-2f90-40df-a66e-141fb5d9f79f-kube-api-access-dqms6\") pod \"nova-cell0-conductor-0\" (UID: \"5f15bdce-2f90-40df-a66e-141fb5d9f79f\") " pod="openstack/nova-cell0-conductor-0" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.708324 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f15bdce-2f90-40df-a66e-141fb5d9f79f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5f15bdce-2f90-40df-a66e-141fb5d9f79f\") " pod="openstack/nova-cell0-conductor-0" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.708456 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f15bdce-2f90-40df-a66e-141fb5d9f79f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5f15bdce-2f90-40df-a66e-141fb5d9f79f\") " pod="openstack/nova-cell0-conductor-0" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.809423 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqms6\" (UniqueName: \"kubernetes.io/projected/5f15bdce-2f90-40df-a66e-141fb5d9f79f-kube-api-access-dqms6\") pod \"nova-cell0-conductor-0\" (UID: \"5f15bdce-2f90-40df-a66e-141fb5d9f79f\") " pod="openstack/nova-cell0-conductor-0" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.809481 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f15bdce-2f90-40df-a66e-141fb5d9f79f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5f15bdce-2f90-40df-a66e-141fb5d9f79f\") " pod="openstack/nova-cell0-conductor-0" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.809562 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f15bdce-2f90-40df-a66e-141fb5d9f79f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5f15bdce-2f90-40df-a66e-141fb5d9f79f\") " pod="openstack/nova-cell0-conductor-0" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.813833 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f15bdce-2f90-40df-a66e-141fb5d9f79f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5f15bdce-2f90-40df-a66e-141fb5d9f79f\") " pod="openstack/nova-cell0-conductor-0" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.814857 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f15bdce-2f90-40df-a66e-141fb5d9f79f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5f15bdce-2f90-40df-a66e-141fb5d9f79f\") " pod="openstack/nova-cell0-conductor-0" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.825058 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqms6\" (UniqueName: \"kubernetes.io/projected/5f15bdce-2f90-40df-a66e-141fb5d9f79f-kube-api-access-dqms6\") pod \"nova-cell0-conductor-0\" (UID: \"5f15bdce-2f90-40df-a66e-141fb5d9f79f\") " pod="openstack/nova-cell0-conductor-0" Nov 22 05:12:45 crc kubenswrapper[4818]: I1122 05:12:45.948467 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 22 05:12:46 crc kubenswrapper[4818]: I1122 05:12:46.279971 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 22 05:12:46 crc kubenswrapper[4818]: W1122 05:12:46.324194 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f15bdce_2f90_40df_a66e_141fb5d9f79f.slice/crio-9af566ce7bfb8c0fad628e5bde2f4eafbfd5249763282b0c34bcfce16ac54225 WatchSource:0}: Error finding container 9af566ce7bfb8c0fad628e5bde2f4eafbfd5249763282b0c34bcfce16ac54225: Status 404 returned error can't find the container with id 9af566ce7bfb8c0fad628e5bde2f4eafbfd5249763282b0c34bcfce16ac54225 Nov 22 05:12:46 crc kubenswrapper[4818]: I1122 05:12:46.481850 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5f15bdce-2f90-40df-a66e-141fb5d9f79f","Type":"ContainerStarted","Data":"de8572c7e9142adfb71f02adecf775a6e53dd6d727f95b911e88ac83e04a33b8"} Nov 22 05:12:46 crc kubenswrapper[4818]: I1122 05:12:46.482156 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 22 05:12:46 crc kubenswrapper[4818]: I1122 05:12:46.482167 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5f15bdce-2f90-40df-a66e-141fb5d9f79f","Type":"ContainerStarted","Data":"9af566ce7bfb8c0fad628e5bde2f4eafbfd5249763282b0c34bcfce16ac54225"} Nov 22 05:12:46 crc kubenswrapper[4818]: I1122 05:12:46.498932 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.498912195 podStartE2EDuration="1.498912195s" podCreationTimestamp="2025-11-22 05:12:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:12:46.496707607 +0000 UTC m=+1519.071124134" watchObservedRunningTime="2025-11-22 05:12:46.498912195 +0000 UTC m=+1519.073328732" Nov 22 05:12:52 crc kubenswrapper[4818]: I1122 05:12:52.634936 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 22 05:12:55 crc kubenswrapper[4818]: I1122 05:12:55.974372 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.481344 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-f9mzx"] Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.483424 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-f9mzx" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.486208 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.486344 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.507235 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-f9mzx"] Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.653211 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7a2b652-b72b-4cd4-85cb-37e996f5238c-scripts\") pod \"nova-cell0-cell-mapping-f9mzx\" (UID: \"c7a2b652-b72b-4cd4-85cb-37e996f5238c\") " pod="openstack/nova-cell0-cell-mapping-f9mzx" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.653306 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7a2b652-b72b-4cd4-85cb-37e996f5238c-config-data\") pod \"nova-cell0-cell-mapping-f9mzx\" (UID: \"c7a2b652-b72b-4cd4-85cb-37e996f5238c\") " pod="openstack/nova-cell0-cell-mapping-f9mzx" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.653362 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7a2b652-b72b-4cd4-85cb-37e996f5238c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-f9mzx\" (UID: \"c7a2b652-b72b-4cd4-85cb-37e996f5238c\") " pod="openstack/nova-cell0-cell-mapping-f9mzx" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.653407 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpq85\" (UniqueName: \"kubernetes.io/projected/c7a2b652-b72b-4cd4-85cb-37e996f5238c-kube-api-access-wpq85\") pod \"nova-cell0-cell-mapping-f9mzx\" (UID: \"c7a2b652-b72b-4cd4-85cb-37e996f5238c\") " pod="openstack/nova-cell0-cell-mapping-f9mzx" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.673673 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.674958 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.677659 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.682979 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.684805 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.692033 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.698492 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.712356 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.758469 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpq85\" (UniqueName: \"kubernetes.io/projected/c7a2b652-b72b-4cd4-85cb-37e996f5238c-kube-api-access-wpq85\") pod \"nova-cell0-cell-mapping-f9mzx\" (UID: \"c7a2b652-b72b-4cd4-85cb-37e996f5238c\") " pod="openstack/nova-cell0-cell-mapping-f9mzx" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.758810 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7a2b652-b72b-4cd4-85cb-37e996f5238c-scripts\") pod \"nova-cell0-cell-mapping-f9mzx\" (UID: \"c7a2b652-b72b-4cd4-85cb-37e996f5238c\") " pod="openstack/nova-cell0-cell-mapping-f9mzx" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.759332 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7a2b652-b72b-4cd4-85cb-37e996f5238c-config-data\") pod \"nova-cell0-cell-mapping-f9mzx\" (UID: \"c7a2b652-b72b-4cd4-85cb-37e996f5238c\") " pod="openstack/nova-cell0-cell-mapping-f9mzx" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.759421 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7a2b652-b72b-4cd4-85cb-37e996f5238c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-f9mzx\" (UID: \"c7a2b652-b72b-4cd4-85cb-37e996f5238c\") " pod="openstack/nova-cell0-cell-mapping-f9mzx" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.767048 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7a2b652-b72b-4cd4-85cb-37e996f5238c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-f9mzx\" (UID: \"c7a2b652-b72b-4cd4-85cb-37e996f5238c\") " pod="openstack/nova-cell0-cell-mapping-f9mzx" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.786372 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7a2b652-b72b-4cd4-85cb-37e996f5238c-scripts\") pod \"nova-cell0-cell-mapping-f9mzx\" (UID: \"c7a2b652-b72b-4cd4-85cb-37e996f5238c\") " pod="openstack/nova-cell0-cell-mapping-f9mzx" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.789878 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7a2b652-b72b-4cd4-85cb-37e996f5238c-config-data\") pod \"nova-cell0-cell-mapping-f9mzx\" (UID: \"c7a2b652-b72b-4cd4-85cb-37e996f5238c\") " pod="openstack/nova-cell0-cell-mapping-f9mzx" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.803552 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpq85\" (UniqueName: \"kubernetes.io/projected/c7a2b652-b72b-4cd4-85cb-37e996f5238c-kube-api-access-wpq85\") pod \"nova-cell0-cell-mapping-f9mzx\" (UID: \"c7a2b652-b72b-4cd4-85cb-37e996f5238c\") " pod="openstack/nova-cell0-cell-mapping-f9mzx" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.826914 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-f9mzx" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.831953 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.834017 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.846969 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.862040 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a78c342-22c9-41f0-9157-9ca44dcae82f-config-data\") pod \"nova-scheduler-0\" (UID: \"5a78c342-22c9-41f0-9157-9ca44dcae82f\") " pod="openstack/nova-scheduler-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.862103 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3df5198-1e08-4baf-811f-493f782a8ed1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b3df5198-1e08-4baf-811f-493f782a8ed1\") " pod="openstack/nova-api-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.862155 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a78c342-22c9-41f0-9157-9ca44dcae82f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5a78c342-22c9-41f0-9157-9ca44dcae82f\") " pod="openstack/nova-scheduler-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.862214 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3df5198-1e08-4baf-811f-493f782a8ed1-logs\") pod \"nova-api-0\" (UID: \"b3df5198-1e08-4baf-811f-493f782a8ed1\") " pod="openstack/nova-api-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.862358 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3df5198-1e08-4baf-811f-493f782a8ed1-config-data\") pod \"nova-api-0\" (UID: \"b3df5198-1e08-4baf-811f-493f782a8ed1\") " pod="openstack/nova-api-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.862384 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b95kf\" (UniqueName: \"kubernetes.io/projected/b3df5198-1e08-4baf-811f-493f782a8ed1-kube-api-access-b95kf\") pod \"nova-api-0\" (UID: \"b3df5198-1e08-4baf-811f-493f782a8ed1\") " pod="openstack/nova-api-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.862440 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmgrd\" (UniqueName: \"kubernetes.io/projected/5a78c342-22c9-41f0-9157-9ca44dcae82f-kube-api-access-fmgrd\") pod \"nova-scheduler-0\" (UID: \"5a78c342-22c9-41f0-9157-9ca44dcae82f\") " pod="openstack/nova-scheduler-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.881238 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.904730 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.906446 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.908596 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.961631 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.963992 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb1ca6ba-4bdf-44d4-8478-59d41abec5d7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb1ca6ba-4bdf-44d4-8478-59d41abec5d7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.964036 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a78c342-22c9-41f0-9157-9ca44dcae82f-config-data\") pod \"nova-scheduler-0\" (UID: \"5a78c342-22c9-41f0-9157-9ca44dcae82f\") " pod="openstack/nova-scheduler-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.964063 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3df5198-1e08-4baf-811f-493f782a8ed1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b3df5198-1e08-4baf-811f-493f782a8ed1\") " pod="openstack/nova-api-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.964098 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a78c342-22c9-41f0-9157-9ca44dcae82f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5a78c342-22c9-41f0-9157-9ca44dcae82f\") " pod="openstack/nova-scheduler-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.964131 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb1ca6ba-4bdf-44d4-8478-59d41abec5d7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb1ca6ba-4bdf-44d4-8478-59d41abec5d7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.964151 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3df5198-1e08-4baf-811f-493f782a8ed1-logs\") pod \"nova-api-0\" (UID: \"b3df5198-1e08-4baf-811f-493f782a8ed1\") " pod="openstack/nova-api-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.964176 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbpnh\" (UniqueName: \"kubernetes.io/projected/cb1ca6ba-4bdf-44d4-8478-59d41abec5d7-kube-api-access-qbpnh\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb1ca6ba-4bdf-44d4-8478-59d41abec5d7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.964248 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3df5198-1e08-4baf-811f-493f782a8ed1-config-data\") pod \"nova-api-0\" (UID: \"b3df5198-1e08-4baf-811f-493f782a8ed1\") " pod="openstack/nova-api-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.964280 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b95kf\" (UniqueName: \"kubernetes.io/projected/b3df5198-1e08-4baf-811f-493f782a8ed1-kube-api-access-b95kf\") pod \"nova-api-0\" (UID: \"b3df5198-1e08-4baf-811f-493f782a8ed1\") " pod="openstack/nova-api-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.964312 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmgrd\" (UniqueName: \"kubernetes.io/projected/5a78c342-22c9-41f0-9157-9ca44dcae82f-kube-api-access-fmgrd\") pod \"nova-scheduler-0\" (UID: \"5a78c342-22c9-41f0-9157-9ca44dcae82f\") " pod="openstack/nova-scheduler-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.966768 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3df5198-1e08-4baf-811f-493f782a8ed1-logs\") pod \"nova-api-0\" (UID: \"b3df5198-1e08-4baf-811f-493f782a8ed1\") " pod="openstack/nova-api-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.977526 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a78c342-22c9-41f0-9157-9ca44dcae82f-config-data\") pod \"nova-scheduler-0\" (UID: \"5a78c342-22c9-41f0-9157-9ca44dcae82f\") " pod="openstack/nova-scheduler-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.977627 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3df5198-1e08-4baf-811f-493f782a8ed1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b3df5198-1e08-4baf-811f-493f782a8ed1\") " pod="openstack/nova-api-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.984722 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-wl6kj"] Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.986657 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.990032 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3df5198-1e08-4baf-811f-493f782a8ed1-config-data\") pod \"nova-api-0\" (UID: \"b3df5198-1e08-4baf-811f-493f782a8ed1\") " pod="openstack/nova-api-0" Nov 22 05:12:56 crc kubenswrapper[4818]: I1122 05:12:56.997031 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-wl6kj"] Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.000456 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a78c342-22c9-41f0-9157-9ca44dcae82f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5a78c342-22c9-41f0-9157-9ca44dcae82f\") " pod="openstack/nova-scheduler-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.004243 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b95kf\" (UniqueName: \"kubernetes.io/projected/b3df5198-1e08-4baf-811f-493f782a8ed1-kube-api-access-b95kf\") pod \"nova-api-0\" (UID: \"b3df5198-1e08-4baf-811f-493f782a8ed1\") " pod="openstack/nova-api-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.005284 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmgrd\" (UniqueName: \"kubernetes.io/projected/5a78c342-22c9-41f0-9157-9ca44dcae82f-kube-api-access-fmgrd\") pod \"nova-scheduler-0\" (UID: \"5a78c342-22c9-41f0-9157-9ca44dcae82f\") " pod="openstack/nova-scheduler-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.006720 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.065595 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb1ca6ba-4bdf-44d4-8478-59d41abec5d7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb1ca6ba-4bdf-44d4-8478-59d41abec5d7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.065889 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbpnh\" (UniqueName: \"kubernetes.io/projected/cb1ca6ba-4bdf-44d4-8478-59d41abec5d7-kube-api-access-qbpnh\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb1ca6ba-4bdf-44d4-8478-59d41abec5d7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.065935 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncx9g\" (UniqueName: \"kubernetes.io/projected/e78732aa-3735-4262-bf33-2726fd0a6ff1-kube-api-access-ncx9g\") pod \"nova-metadata-0\" (UID: \"e78732aa-3735-4262-bf33-2726fd0a6ff1\") " pod="openstack/nova-metadata-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.066030 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb1ca6ba-4bdf-44d4-8478-59d41abec5d7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb1ca6ba-4bdf-44d4-8478-59d41abec5d7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.066062 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e78732aa-3735-4262-bf33-2726fd0a6ff1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e78732aa-3735-4262-bf33-2726fd0a6ff1\") " pod="openstack/nova-metadata-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.066078 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e78732aa-3735-4262-bf33-2726fd0a6ff1-logs\") pod \"nova-metadata-0\" (UID: \"e78732aa-3735-4262-bf33-2726fd0a6ff1\") " pod="openstack/nova-metadata-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.066097 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e78732aa-3735-4262-bf33-2726fd0a6ff1-config-data\") pod \"nova-metadata-0\" (UID: \"e78732aa-3735-4262-bf33-2726fd0a6ff1\") " pod="openstack/nova-metadata-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.070290 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb1ca6ba-4bdf-44d4-8478-59d41abec5d7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb1ca6ba-4bdf-44d4-8478-59d41abec5d7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.072911 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb1ca6ba-4bdf-44d4-8478-59d41abec5d7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb1ca6ba-4bdf-44d4-8478-59d41abec5d7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.084525 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbpnh\" (UniqueName: \"kubernetes.io/projected/cb1ca6ba-4bdf-44d4-8478-59d41abec5d7-kube-api-access-qbpnh\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb1ca6ba-4bdf-44d4-8478-59d41abec5d7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.167879 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-ovsdbserver-nb\") pod \"dnsmasq-dns-566b5b7845-wl6kj\" (UID: \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\") " pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.168084 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e78732aa-3735-4262-bf33-2726fd0a6ff1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e78732aa-3735-4262-bf33-2726fd0a6ff1\") " pod="openstack/nova-metadata-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.168108 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e78732aa-3735-4262-bf33-2726fd0a6ff1-logs\") pod \"nova-metadata-0\" (UID: \"e78732aa-3735-4262-bf33-2726fd0a6ff1\") " pod="openstack/nova-metadata-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.168142 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e78732aa-3735-4262-bf33-2726fd0a6ff1-config-data\") pod \"nova-metadata-0\" (UID: \"e78732aa-3735-4262-bf33-2726fd0a6ff1\") " pod="openstack/nova-metadata-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.168172 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-config\") pod \"dnsmasq-dns-566b5b7845-wl6kj\" (UID: \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\") " pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.168202 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-ovsdbserver-sb\") pod \"dnsmasq-dns-566b5b7845-wl6kj\" (UID: \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\") " pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.168295 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-dns-svc\") pod \"dnsmasq-dns-566b5b7845-wl6kj\" (UID: \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\") " pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.168346 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncx9g\" (UniqueName: \"kubernetes.io/projected/e78732aa-3735-4262-bf33-2726fd0a6ff1-kube-api-access-ncx9g\") pod \"nova-metadata-0\" (UID: \"e78732aa-3735-4262-bf33-2726fd0a6ff1\") " pod="openstack/nova-metadata-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.168407 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f4lc\" (UniqueName: \"kubernetes.io/projected/34665b87-c9bb-4b11-9406-57bc79ed2e0c-kube-api-access-7f4lc\") pod \"dnsmasq-dns-566b5b7845-wl6kj\" (UID: \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\") " pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.171514 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e78732aa-3735-4262-bf33-2726fd0a6ff1-logs\") pod \"nova-metadata-0\" (UID: \"e78732aa-3735-4262-bf33-2726fd0a6ff1\") " pod="openstack/nova-metadata-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.174144 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e78732aa-3735-4262-bf33-2726fd0a6ff1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e78732aa-3735-4262-bf33-2726fd0a6ff1\") " pod="openstack/nova-metadata-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.179933 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e78732aa-3735-4262-bf33-2726fd0a6ff1-config-data\") pod \"nova-metadata-0\" (UID: \"e78732aa-3735-4262-bf33-2726fd0a6ff1\") " pod="openstack/nova-metadata-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.187854 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncx9g\" (UniqueName: \"kubernetes.io/projected/e78732aa-3735-4262-bf33-2726fd0a6ff1-kube-api-access-ncx9g\") pod \"nova-metadata-0\" (UID: \"e78732aa-3735-4262-bf33-2726fd0a6ff1\") " pod="openstack/nova-metadata-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.270707 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f4lc\" (UniqueName: \"kubernetes.io/projected/34665b87-c9bb-4b11-9406-57bc79ed2e0c-kube-api-access-7f4lc\") pod \"dnsmasq-dns-566b5b7845-wl6kj\" (UID: \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\") " pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.270798 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-ovsdbserver-nb\") pod \"dnsmasq-dns-566b5b7845-wl6kj\" (UID: \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\") " pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.270861 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-config\") pod \"dnsmasq-dns-566b5b7845-wl6kj\" (UID: \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\") " pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.270914 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-ovsdbserver-sb\") pod \"dnsmasq-dns-566b5b7845-wl6kj\" (UID: \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\") " pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.271871 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-config\") pod \"dnsmasq-dns-566b5b7845-wl6kj\" (UID: \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\") " pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.271893 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-ovsdbserver-sb\") pod \"dnsmasq-dns-566b5b7845-wl6kj\" (UID: \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\") " pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.270956 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-dns-svc\") pod \"dnsmasq-dns-566b5b7845-wl6kj\" (UID: \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\") " pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.272208 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-dns-svc\") pod \"dnsmasq-dns-566b5b7845-wl6kj\" (UID: \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\") " pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.272407 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-ovsdbserver-nb\") pod \"dnsmasq-dns-566b5b7845-wl6kj\" (UID: \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\") " pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.290679 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.295113 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f4lc\" (UniqueName: \"kubernetes.io/projected/34665b87-c9bb-4b11-9406-57bc79ed2e0c-kube-api-access-7f4lc\") pod \"dnsmasq-dns-566b5b7845-wl6kj\" (UID: \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\") " pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.320033 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.333458 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.352730 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.413295 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-f9mzx"] Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.552166 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.569388 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tfdvb"] Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.570859 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tfdvb" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.573779 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.573876 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.577588 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tfdvb"] Nov 22 05:12:57 crc kubenswrapper[4818]: W1122 05:12:57.590960 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3df5198_1e08_4baf_811f_493f782a8ed1.slice/crio-db0e78bb921c92dc07d69e0574a9c0d9d9e0a7dc88fe3f8148c2a02ccc2c1e87 WatchSource:0}: Error finding container db0e78bb921c92dc07d69e0574a9c0d9d9e0a7dc88fe3f8148c2a02ccc2c1e87: Status 404 returned error can't find the container with id db0e78bb921c92dc07d69e0574a9c0d9d9e0a7dc88fe3f8148c2a02ccc2c1e87 Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.594637 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.681834 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-config-data\") pod \"nova-cell1-conductor-db-sync-tfdvb\" (UID: \"10dfb213-2885-405b-8c78-78363efec183\") " pod="openstack/nova-cell1-conductor-db-sync-tfdvb" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.681909 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tfdvb\" (UID: \"10dfb213-2885-405b-8c78-78363efec183\") " pod="openstack/nova-cell1-conductor-db-sync-tfdvb" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.682033 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gt2w\" (UniqueName: \"kubernetes.io/projected/10dfb213-2885-405b-8c78-78363efec183-kube-api-access-5gt2w\") pod \"nova-cell1-conductor-db-sync-tfdvb\" (UID: \"10dfb213-2885-405b-8c78-78363efec183\") " pod="openstack/nova-cell1-conductor-db-sync-tfdvb" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.682066 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-scripts\") pod \"nova-cell1-conductor-db-sync-tfdvb\" (UID: \"10dfb213-2885-405b-8c78-78363efec183\") " pod="openstack/nova-cell1-conductor-db-sync-tfdvb" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.727454 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 05:12:57 crc kubenswrapper[4818]: W1122 05:12:57.731060 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a78c342_22c9_41f0_9157_9ca44dcae82f.slice/crio-d90344a81b4102caa0b1e3859a0722007dacd6c5bce61ebf9548678bff4069bd WatchSource:0}: Error finding container d90344a81b4102caa0b1e3859a0722007dacd6c5bce61ebf9548678bff4069bd: Status 404 returned error can't find the container with id d90344a81b4102caa0b1e3859a0722007dacd6c5bce61ebf9548678bff4069bd Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.783418 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-config-data\") pod \"nova-cell1-conductor-db-sync-tfdvb\" (UID: \"10dfb213-2885-405b-8c78-78363efec183\") " pod="openstack/nova-cell1-conductor-db-sync-tfdvb" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.783507 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tfdvb\" (UID: \"10dfb213-2885-405b-8c78-78363efec183\") " pod="openstack/nova-cell1-conductor-db-sync-tfdvb" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.783614 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gt2w\" (UniqueName: \"kubernetes.io/projected/10dfb213-2885-405b-8c78-78363efec183-kube-api-access-5gt2w\") pod \"nova-cell1-conductor-db-sync-tfdvb\" (UID: \"10dfb213-2885-405b-8c78-78363efec183\") " pod="openstack/nova-cell1-conductor-db-sync-tfdvb" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.783653 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-scripts\") pod \"nova-cell1-conductor-db-sync-tfdvb\" (UID: \"10dfb213-2885-405b-8c78-78363efec183\") " pod="openstack/nova-cell1-conductor-db-sync-tfdvb" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.789037 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-scripts\") pod \"nova-cell1-conductor-db-sync-tfdvb\" (UID: \"10dfb213-2885-405b-8c78-78363efec183\") " pod="openstack/nova-cell1-conductor-db-sync-tfdvb" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.789110 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-config-data\") pod \"nova-cell1-conductor-db-sync-tfdvb\" (UID: \"10dfb213-2885-405b-8c78-78363efec183\") " pod="openstack/nova-cell1-conductor-db-sync-tfdvb" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.799012 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tfdvb\" (UID: \"10dfb213-2885-405b-8c78-78363efec183\") " pod="openstack/nova-cell1-conductor-db-sync-tfdvb" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.807133 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gt2w\" (UniqueName: \"kubernetes.io/projected/10dfb213-2885-405b-8c78-78363efec183-kube-api-access-5gt2w\") pod \"nova-cell1-conductor-db-sync-tfdvb\" (UID: \"10dfb213-2885-405b-8c78-78363efec183\") " pod="openstack/nova-cell1-conductor-db-sync-tfdvb" Nov 22 05:12:57 crc kubenswrapper[4818]: W1122 05:12:57.898496 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb1ca6ba_4bdf_44d4_8478_59d41abec5d7.slice/crio-08401c280cc3e65837f99f61db45e33c90372a256a33426ef9ecdb0fa16a4dd7 WatchSource:0}: Error finding container 08401c280cc3e65837f99f61db45e33c90372a256a33426ef9ecdb0fa16a4dd7: Status 404 returned error can't find the container with id 08401c280cc3e65837f99f61db45e33c90372a256a33426ef9ecdb0fa16a4dd7 Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.900992 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.905843 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5a78c342-22c9-41f0-9157-9ca44dcae82f","Type":"ContainerStarted","Data":"d90344a81b4102caa0b1e3859a0722007dacd6c5bce61ebf9548678bff4069bd"} Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.906711 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b3df5198-1e08-4baf-811f-493f782a8ed1","Type":"ContainerStarted","Data":"db0e78bb921c92dc07d69e0574a9c0d9d9e0a7dc88fe3f8148c2a02ccc2c1e87"} Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.907781 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-f9mzx" event={"ID":"c7a2b652-b72b-4cd4-85cb-37e996f5238c","Type":"ContainerStarted","Data":"3a41077bbf785cb7d8ba47f08356c586c5e42d79788b494a74d45edd2c6ea260"} Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.907821 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-f9mzx" event={"ID":"c7a2b652-b72b-4cd4-85cb-37e996f5238c","Type":"ContainerStarted","Data":"dbf7868660a6b0d72ce82662a893b6dab41b0127e61cd8feb85eb005e0448dcc"} Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.925817 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tfdvb" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.926089 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-f9mzx" podStartSLOduration=1.926070135 podStartE2EDuration="1.926070135s" podCreationTimestamp="2025-11-22 05:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:12:57.918954368 +0000 UTC m=+1530.493370895" watchObservedRunningTime="2025-11-22 05:12:57.926070135 +0000 UTC m=+1530.500486662" Nov 22 05:12:57 crc kubenswrapper[4818]: I1122 05:12:57.997106 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 05:12:58 crc kubenswrapper[4818]: I1122 05:12:58.048194 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-wl6kj"] Nov 22 05:12:58 crc kubenswrapper[4818]: I1122 05:12:58.381014 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tfdvb"] Nov 22 05:12:58 crc kubenswrapper[4818]: W1122 05:12:58.382437 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10dfb213_2885_405b_8c78_78363efec183.slice/crio-dfac1c10ecbf4be3ae77df2eff5811ed0fa0b3dec7527cdffd061d1f32cc1495 WatchSource:0}: Error finding container dfac1c10ecbf4be3ae77df2eff5811ed0fa0b3dec7527cdffd061d1f32cc1495: Status 404 returned error can't find the container with id dfac1c10ecbf4be3ae77df2eff5811ed0fa0b3dec7527cdffd061d1f32cc1495 Nov 22 05:12:58 crc kubenswrapper[4818]: I1122 05:12:58.917919 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e78732aa-3735-4262-bf33-2726fd0a6ff1","Type":"ContainerStarted","Data":"057e8530e64d5501a445e4514710c54a3028346b64cebf7f2245814b4a8725d6"} Nov 22 05:12:58 crc kubenswrapper[4818]: I1122 05:12:58.920908 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tfdvb" event={"ID":"10dfb213-2885-405b-8c78-78363efec183","Type":"ContainerStarted","Data":"4f8a0beee11a41b125ccd8bcedeca4bb9b6a58dc64b4e10d3c2f83b5853f9851"} Nov 22 05:12:58 crc kubenswrapper[4818]: I1122 05:12:58.920968 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tfdvb" event={"ID":"10dfb213-2885-405b-8c78-78363efec183","Type":"ContainerStarted","Data":"dfac1c10ecbf4be3ae77df2eff5811ed0fa0b3dec7527cdffd061d1f32cc1495"} Nov 22 05:12:58 crc kubenswrapper[4818]: I1122 05:12:58.926178 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cb1ca6ba-4bdf-44d4-8478-59d41abec5d7","Type":"ContainerStarted","Data":"08401c280cc3e65837f99f61db45e33c90372a256a33426ef9ecdb0fa16a4dd7"} Nov 22 05:12:58 crc kubenswrapper[4818]: I1122 05:12:58.929826 4818 generic.go:334] "Generic (PLEG): container finished" podID="34665b87-c9bb-4b11-9406-57bc79ed2e0c" containerID="441378f5a899b66f84bdf03159b3460c7aebc5518fd93ee42ae52eea9f2ae661" exitCode=0 Nov 22 05:12:58 crc kubenswrapper[4818]: I1122 05:12:58.929968 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" event={"ID":"34665b87-c9bb-4b11-9406-57bc79ed2e0c","Type":"ContainerDied","Data":"441378f5a899b66f84bdf03159b3460c7aebc5518fd93ee42ae52eea9f2ae661"} Nov 22 05:12:58 crc kubenswrapper[4818]: I1122 05:12:58.930027 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" event={"ID":"34665b87-c9bb-4b11-9406-57bc79ed2e0c","Type":"ContainerStarted","Data":"4f6898013226a2a6e98093db4154ae814ce19b76449da3c797b00ce7b3bdd899"} Nov 22 05:12:58 crc kubenswrapper[4818]: I1122 05:12:58.946242 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-tfdvb" podStartSLOduration=1.946217055 podStartE2EDuration="1.946217055s" podCreationTimestamp="2025-11-22 05:12:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:12:58.936869468 +0000 UTC m=+1531.511285995" watchObservedRunningTime="2025-11-22 05:12:58.946217055 +0000 UTC m=+1531.520633582" Nov 22 05:13:00 crc kubenswrapper[4818]: I1122 05:13:00.305238 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 05:13:00 crc kubenswrapper[4818]: I1122 05:13:00.348001 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 05:13:01 crc kubenswrapper[4818]: I1122 05:13:01.956982 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" event={"ID":"34665b87-c9bb-4b11-9406-57bc79ed2e0c","Type":"ContainerStarted","Data":"3bbe4c52bc939b728f48822e01bec5b824c46735ebe31147ec275e3f4aa58476"} Nov 22 05:13:01 crc kubenswrapper[4818]: I1122 05:13:01.957581 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:13:01 crc kubenswrapper[4818]: I1122 05:13:01.958924 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5a78c342-22c9-41f0-9157-9ca44dcae82f","Type":"ContainerStarted","Data":"125e1c460f3556eca44f4dd4b5178b464cdb2f73230f2728c7faff73312b1ede"} Nov 22 05:13:01 crc kubenswrapper[4818]: I1122 05:13:01.963667 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e78732aa-3735-4262-bf33-2726fd0a6ff1","Type":"ContainerStarted","Data":"7f2b20c7ad72020b50ab18a72c91647f36d316ff9d9b4007722aabac0d1e6b9d"} Nov 22 05:13:01 crc kubenswrapper[4818]: I1122 05:13:01.963707 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e78732aa-3735-4262-bf33-2726fd0a6ff1","Type":"ContainerStarted","Data":"17684c1a5c2c94a609e2c8edff7f798e0d6a65ca7f711fdf7ce896fba696da5e"} Nov 22 05:13:01 crc kubenswrapper[4818]: I1122 05:13:01.963739 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e78732aa-3735-4262-bf33-2726fd0a6ff1" containerName="nova-metadata-log" containerID="cri-o://17684c1a5c2c94a609e2c8edff7f798e0d6a65ca7f711fdf7ce896fba696da5e" gracePeriod=30 Nov 22 05:13:01 crc kubenswrapper[4818]: I1122 05:13:01.963824 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e78732aa-3735-4262-bf33-2726fd0a6ff1" containerName="nova-metadata-metadata" containerID="cri-o://7f2b20c7ad72020b50ab18a72c91647f36d316ff9d9b4007722aabac0d1e6b9d" gracePeriod=30 Nov 22 05:13:01 crc kubenswrapper[4818]: I1122 05:13:01.966572 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b3df5198-1e08-4baf-811f-493f782a8ed1","Type":"ContainerStarted","Data":"9fc067ba8f3205ac4e510123b4beddd9eff6a6fc1b1ffb3cd4d6ec2d9e204d63"} Nov 22 05:13:01 crc kubenswrapper[4818]: I1122 05:13:01.966607 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b3df5198-1e08-4baf-811f-493f782a8ed1","Type":"ContainerStarted","Data":"3b9d7e170741eedb47592033e9a1d7b80e86fb1648481543e959d67c29bd8084"} Nov 22 05:13:01 crc kubenswrapper[4818]: I1122 05:13:01.969526 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cb1ca6ba-4bdf-44d4-8478-59d41abec5d7","Type":"ContainerStarted","Data":"7f6074bbb86e476297d4ead08a31f1c8c1400eed778d33e5b8b10eb527e58d43"} Nov 22 05:13:01 crc kubenswrapper[4818]: I1122 05:13:01.969655 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="cb1ca6ba-4bdf-44d4-8478-59d41abec5d7" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://7f6074bbb86e476297d4ead08a31f1c8c1400eed778d33e5b8b10eb527e58d43" gracePeriod=30 Nov 22 05:13:01 crc kubenswrapper[4818]: I1122 05:13:01.981888 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" podStartSLOduration=5.981872789 podStartE2EDuration="5.981872789s" podCreationTimestamp="2025-11-22 05:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:13:01.978970542 +0000 UTC m=+1534.553387089" watchObservedRunningTime="2025-11-22 05:13:01.981872789 +0000 UTC m=+1534.556289316" Nov 22 05:13:02 crc kubenswrapper[4818]: I1122 05:13:02.004589 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.674254624 podStartE2EDuration="6.004571787s" podCreationTimestamp="2025-11-22 05:12:56 +0000 UTC" firstStartedPulling="2025-11-22 05:12:57.90156352 +0000 UTC m=+1530.475980047" lastFinishedPulling="2025-11-22 05:13:01.231880683 +0000 UTC m=+1533.806297210" observedRunningTime="2025-11-22 05:13:02.003003696 +0000 UTC m=+1534.577420233" watchObservedRunningTime="2025-11-22 05:13:02.004571787 +0000 UTC m=+1534.578988314" Nov 22 05:13:02 crc kubenswrapper[4818]: I1122 05:13:02.041175 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.546858385 podStartE2EDuration="6.041151992s" podCreationTimestamp="2025-11-22 05:12:56 +0000 UTC" firstStartedPulling="2025-11-22 05:12:57.737433482 +0000 UTC m=+1530.311850009" lastFinishedPulling="2025-11-22 05:13:01.231727089 +0000 UTC m=+1533.806143616" observedRunningTime="2025-11-22 05:13:02.021789842 +0000 UTC m=+1534.596206379" watchObservedRunningTime="2025-11-22 05:13:02.041151992 +0000 UTC m=+1534.615568519" Nov 22 05:13:02 crc kubenswrapper[4818]: I1122 05:13:02.041706 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.423217864 podStartE2EDuration="6.041700887s" podCreationTimestamp="2025-11-22 05:12:56 +0000 UTC" firstStartedPulling="2025-11-22 05:12:57.594304957 +0000 UTC m=+1530.168721474" lastFinishedPulling="2025-11-22 05:13:01.21278795 +0000 UTC m=+1533.787204497" observedRunningTime="2025-11-22 05:13:02.035460172 +0000 UTC m=+1534.609876699" watchObservedRunningTime="2025-11-22 05:13:02.041700887 +0000 UTC m=+1534.616117414" Nov 22 05:13:02 crc kubenswrapper[4818]: I1122 05:13:02.064083 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.83453675 podStartE2EDuration="6.064062256s" podCreationTimestamp="2025-11-22 05:12:56 +0000 UTC" firstStartedPulling="2025-11-22 05:12:58.001067943 +0000 UTC m=+1530.575484470" lastFinishedPulling="2025-11-22 05:13:01.230593449 +0000 UTC m=+1533.805009976" observedRunningTime="2025-11-22 05:13:02.058589282 +0000 UTC m=+1534.633005809" watchObservedRunningTime="2025-11-22 05:13:02.064062256 +0000 UTC m=+1534.638478783" Nov 22 05:13:02 crc kubenswrapper[4818]: I1122 05:13:02.301410 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 22 05:13:02 crc kubenswrapper[4818]: I1122 05:13:02.321157 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:02 crc kubenswrapper[4818]: I1122 05:13:02.334424 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 22 05:13:02 crc kubenswrapper[4818]: I1122 05:13:02.334474 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 22 05:13:02 crc kubenswrapper[4818]: I1122 05:13:02.984660 4818 generic.go:334] "Generic (PLEG): container finished" podID="e78732aa-3735-4262-bf33-2726fd0a6ff1" containerID="17684c1a5c2c94a609e2c8edff7f798e0d6a65ca7f711fdf7ce896fba696da5e" exitCode=143 Nov 22 05:13:02 crc kubenswrapper[4818]: I1122 05:13:02.985687 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e78732aa-3735-4262-bf33-2726fd0a6ff1","Type":"ContainerDied","Data":"17684c1a5c2c94a609e2c8edff7f798e0d6a65ca7f711fdf7ce896fba696da5e"} Nov 22 05:13:05 crc kubenswrapper[4818]: I1122 05:13:05.008510 4818 generic.go:334] "Generic (PLEG): container finished" podID="c7a2b652-b72b-4cd4-85cb-37e996f5238c" containerID="3a41077bbf785cb7d8ba47f08356c586c5e42d79788b494a74d45edd2c6ea260" exitCode=0 Nov 22 05:13:05 crc kubenswrapper[4818]: I1122 05:13:05.008699 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-f9mzx" event={"ID":"c7a2b652-b72b-4cd4-85cb-37e996f5238c","Type":"ContainerDied","Data":"3a41077bbf785cb7d8ba47f08356c586c5e42d79788b494a74d45edd2c6ea260"} Nov 22 05:13:06 crc kubenswrapper[4818]: I1122 05:13:06.362282 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-f9mzx" Nov 22 05:13:06 crc kubenswrapper[4818]: I1122 05:13:06.478653 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7a2b652-b72b-4cd4-85cb-37e996f5238c-combined-ca-bundle\") pod \"c7a2b652-b72b-4cd4-85cb-37e996f5238c\" (UID: \"c7a2b652-b72b-4cd4-85cb-37e996f5238c\") " Nov 22 05:13:06 crc kubenswrapper[4818]: I1122 05:13:06.478954 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpq85\" (UniqueName: \"kubernetes.io/projected/c7a2b652-b72b-4cd4-85cb-37e996f5238c-kube-api-access-wpq85\") pod \"c7a2b652-b72b-4cd4-85cb-37e996f5238c\" (UID: \"c7a2b652-b72b-4cd4-85cb-37e996f5238c\") " Nov 22 05:13:06 crc kubenswrapper[4818]: I1122 05:13:06.479117 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7a2b652-b72b-4cd4-85cb-37e996f5238c-config-data\") pod \"c7a2b652-b72b-4cd4-85cb-37e996f5238c\" (UID: \"c7a2b652-b72b-4cd4-85cb-37e996f5238c\") " Nov 22 05:13:06 crc kubenswrapper[4818]: I1122 05:13:06.479164 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7a2b652-b72b-4cd4-85cb-37e996f5238c-scripts\") pod \"c7a2b652-b72b-4cd4-85cb-37e996f5238c\" (UID: \"c7a2b652-b72b-4cd4-85cb-37e996f5238c\") " Nov 22 05:13:06 crc kubenswrapper[4818]: I1122 05:13:06.484818 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7a2b652-b72b-4cd4-85cb-37e996f5238c-kube-api-access-wpq85" (OuterVolumeSpecName: "kube-api-access-wpq85") pod "c7a2b652-b72b-4cd4-85cb-37e996f5238c" (UID: "c7a2b652-b72b-4cd4-85cb-37e996f5238c"). InnerVolumeSpecName "kube-api-access-wpq85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:13:06 crc kubenswrapper[4818]: I1122 05:13:06.484897 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7a2b652-b72b-4cd4-85cb-37e996f5238c-scripts" (OuterVolumeSpecName: "scripts") pod "c7a2b652-b72b-4cd4-85cb-37e996f5238c" (UID: "c7a2b652-b72b-4cd4-85cb-37e996f5238c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:06 crc kubenswrapper[4818]: I1122 05:13:06.504507 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7a2b652-b72b-4cd4-85cb-37e996f5238c-config-data" (OuterVolumeSpecName: "config-data") pod "c7a2b652-b72b-4cd4-85cb-37e996f5238c" (UID: "c7a2b652-b72b-4cd4-85cb-37e996f5238c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:06 crc kubenswrapper[4818]: I1122 05:13:06.507243 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7a2b652-b72b-4cd4-85cb-37e996f5238c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c7a2b652-b72b-4cd4-85cb-37e996f5238c" (UID: "c7a2b652-b72b-4cd4-85cb-37e996f5238c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:06 crc kubenswrapper[4818]: I1122 05:13:06.582037 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7a2b652-b72b-4cd4-85cb-37e996f5238c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:06 crc kubenswrapper[4818]: I1122 05:13:06.582071 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpq85\" (UniqueName: \"kubernetes.io/projected/c7a2b652-b72b-4cd4-85cb-37e996f5238c-kube-api-access-wpq85\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:06 crc kubenswrapper[4818]: I1122 05:13:06.582083 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7a2b652-b72b-4cd4-85cb-37e996f5238c-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:06 crc kubenswrapper[4818]: I1122 05:13:06.582091 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7a2b652-b72b-4cd4-85cb-37e996f5238c-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:07 crc kubenswrapper[4818]: I1122 05:13:07.008650 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 05:13:07 crc kubenswrapper[4818]: I1122 05:13:07.008719 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 05:13:07 crc kubenswrapper[4818]: I1122 05:13:07.033288 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-f9mzx" event={"ID":"c7a2b652-b72b-4cd4-85cb-37e996f5238c","Type":"ContainerDied","Data":"dbf7868660a6b0d72ce82662a893b6dab41b0127e61cd8feb85eb005e0448dcc"} Nov 22 05:13:07 crc kubenswrapper[4818]: I1122 05:13:07.034383 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbf7868660a6b0d72ce82662a893b6dab41b0127e61cd8feb85eb005e0448dcc" Nov 22 05:13:07 crc kubenswrapper[4818]: I1122 05:13:07.033359 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-f9mzx" Nov 22 05:13:07 crc kubenswrapper[4818]: I1122 05:13:07.238460 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 22 05:13:07 crc kubenswrapper[4818]: I1122 05:13:07.238769 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b3df5198-1e08-4baf-811f-493f782a8ed1" containerName="nova-api-log" containerID="cri-o://3b9d7e170741eedb47592033e9a1d7b80e86fb1648481543e959d67c29bd8084" gracePeriod=30 Nov 22 05:13:07 crc kubenswrapper[4818]: I1122 05:13:07.238854 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b3df5198-1e08-4baf-811f-493f782a8ed1" containerName="nova-api-api" containerID="cri-o://9fc067ba8f3205ac4e510123b4beddd9eff6a6fc1b1ffb3cd4d6ec2d9e204d63" gracePeriod=30 Nov 22 05:13:07 crc kubenswrapper[4818]: I1122 05:13:07.250640 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b3df5198-1e08-4baf-811f-493f782a8ed1" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.170:8774/\": EOF" Nov 22 05:13:07 crc kubenswrapper[4818]: I1122 05:13:07.251452 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b3df5198-1e08-4baf-811f-493f782a8ed1" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.170:8774/\": EOF" Nov 22 05:13:07 crc kubenswrapper[4818]: I1122 05:13:07.252123 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 05:13:07 crc kubenswrapper[4818]: I1122 05:13:07.252402 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="5a78c342-22c9-41f0-9157-9ca44dcae82f" containerName="nova-scheduler-scheduler" containerID="cri-o://125e1c460f3556eca44f4dd4b5178b464cdb2f73230f2728c7faff73312b1ede" gracePeriod=30 Nov 22 05:13:07 crc kubenswrapper[4818]: I1122 05:13:07.358521 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:13:07 crc kubenswrapper[4818]: I1122 05:13:07.409018 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-dh7x5"] Nov 22 05:13:07 crc kubenswrapper[4818]: I1122 05:13:07.409597 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" podUID="a527f913-2977-419c-9a4c-25b4c1729025" containerName="dnsmasq-dns" containerID="cri-o://dd79d8485fd85c4d0789a9fcddee8a153cd2a022fc0dce7e1cadc5e4a9cbaa05" gracePeriod=10 Nov 22 05:13:07 crc kubenswrapper[4818]: I1122 05:13:07.936532 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.043153 4818 generic.go:334] "Generic (PLEG): container finished" podID="a527f913-2977-419c-9a4c-25b4c1729025" containerID="dd79d8485fd85c4d0789a9fcddee8a153cd2a022fc0dce7e1cadc5e4a9cbaa05" exitCode=0 Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.043200 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" event={"ID":"a527f913-2977-419c-9a4c-25b4c1729025","Type":"ContainerDied","Data":"dd79d8485fd85c4d0789a9fcddee8a153cd2a022fc0dce7e1cadc5e4a9cbaa05"} Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.043264 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" event={"ID":"a527f913-2977-419c-9a4c-25b4c1729025","Type":"ContainerDied","Data":"7c370fad63c7f518851650c725bc3157cdcbee262d848a6db0078aa6dd1b56be"} Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.043277 4818 scope.go:117] "RemoveContainer" containerID="dd79d8485fd85c4d0789a9fcddee8a153cd2a022fc0dce7e1cadc5e4a9cbaa05" Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.043285 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-dh7x5" Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.044941 4818 generic.go:334] "Generic (PLEG): container finished" podID="b3df5198-1e08-4baf-811f-493f782a8ed1" containerID="3b9d7e170741eedb47592033e9a1d7b80e86fb1648481543e959d67c29bd8084" exitCode=143 Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.044990 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b3df5198-1e08-4baf-811f-493f782a8ed1","Type":"ContainerDied","Data":"3b9d7e170741eedb47592033e9a1d7b80e86fb1648481543e959d67c29bd8084"} Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.063243 4818 scope.go:117] "RemoveContainer" containerID="a85c8dbbfe43d1f5977724516652d8cd47242546de587eac7870ab7585c6c0f7" Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.087632 4818 scope.go:117] "RemoveContainer" containerID="dd79d8485fd85c4d0789a9fcddee8a153cd2a022fc0dce7e1cadc5e4a9cbaa05" Nov 22 05:13:08 crc kubenswrapper[4818]: E1122 05:13:08.088407 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd79d8485fd85c4d0789a9fcddee8a153cd2a022fc0dce7e1cadc5e4a9cbaa05\": container with ID starting with dd79d8485fd85c4d0789a9fcddee8a153cd2a022fc0dce7e1cadc5e4a9cbaa05 not found: ID does not exist" containerID="dd79d8485fd85c4d0789a9fcddee8a153cd2a022fc0dce7e1cadc5e4a9cbaa05" Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.088453 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd79d8485fd85c4d0789a9fcddee8a153cd2a022fc0dce7e1cadc5e4a9cbaa05"} err="failed to get container status \"dd79d8485fd85c4d0789a9fcddee8a153cd2a022fc0dce7e1cadc5e4a9cbaa05\": rpc error: code = NotFound desc = could not find container \"dd79d8485fd85c4d0789a9fcddee8a153cd2a022fc0dce7e1cadc5e4a9cbaa05\": container with ID starting with dd79d8485fd85c4d0789a9fcddee8a153cd2a022fc0dce7e1cadc5e4a9cbaa05 not found: ID does not exist" Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.088484 4818 scope.go:117] "RemoveContainer" containerID="a85c8dbbfe43d1f5977724516652d8cd47242546de587eac7870ab7585c6c0f7" Nov 22 05:13:08 crc kubenswrapper[4818]: E1122 05:13:08.088818 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a85c8dbbfe43d1f5977724516652d8cd47242546de587eac7870ab7585c6c0f7\": container with ID starting with a85c8dbbfe43d1f5977724516652d8cd47242546de587eac7870ab7585c6c0f7 not found: ID does not exist" containerID="a85c8dbbfe43d1f5977724516652d8cd47242546de587eac7870ab7585c6c0f7" Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.088850 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a85c8dbbfe43d1f5977724516652d8cd47242546de587eac7870ab7585c6c0f7"} err="failed to get container status \"a85c8dbbfe43d1f5977724516652d8cd47242546de587eac7870ab7585c6c0f7\": rpc error: code = NotFound desc = could not find container \"a85c8dbbfe43d1f5977724516652d8cd47242546de587eac7870ab7585c6c0f7\": container with ID starting with a85c8dbbfe43d1f5977724516652d8cd47242546de587eac7870ab7585c6c0f7 not found: ID does not exist" Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.116151 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-dns-svc\") pod \"a527f913-2977-419c-9a4c-25b4c1729025\" (UID: \"a527f913-2977-419c-9a4c-25b4c1729025\") " Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.116295 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-ovsdbserver-sb\") pod \"a527f913-2977-419c-9a4c-25b4c1729025\" (UID: \"a527f913-2977-419c-9a4c-25b4c1729025\") " Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.116322 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-ovsdbserver-nb\") pod \"a527f913-2977-419c-9a4c-25b4c1729025\" (UID: \"a527f913-2977-419c-9a4c-25b4c1729025\") " Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.116358 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-config\") pod \"a527f913-2977-419c-9a4c-25b4c1729025\" (UID: \"a527f913-2977-419c-9a4c-25b4c1729025\") " Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.116401 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v76vm\" (UniqueName: \"kubernetes.io/projected/a527f913-2977-419c-9a4c-25b4c1729025-kube-api-access-v76vm\") pod \"a527f913-2977-419c-9a4c-25b4c1729025\" (UID: \"a527f913-2977-419c-9a4c-25b4c1729025\") " Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.122467 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a527f913-2977-419c-9a4c-25b4c1729025-kube-api-access-v76vm" (OuterVolumeSpecName: "kube-api-access-v76vm") pod "a527f913-2977-419c-9a4c-25b4c1729025" (UID: "a527f913-2977-419c-9a4c-25b4c1729025"). InnerVolumeSpecName "kube-api-access-v76vm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.165056 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a527f913-2977-419c-9a4c-25b4c1729025" (UID: "a527f913-2977-419c-9a4c-25b4c1729025"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.175446 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-config" (OuterVolumeSpecName: "config") pod "a527f913-2977-419c-9a4c-25b4c1729025" (UID: "a527f913-2977-419c-9a4c-25b4c1729025"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.180541 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a527f913-2977-419c-9a4c-25b4c1729025" (UID: "a527f913-2977-419c-9a4c-25b4c1729025"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.180880 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a527f913-2977-419c-9a4c-25b4c1729025" (UID: "a527f913-2977-419c-9a4c-25b4c1729025"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.219343 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v76vm\" (UniqueName: \"kubernetes.io/projected/a527f913-2977-419c-9a4c-25b4c1729025-kube-api-access-v76vm\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.219379 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.219391 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.219401 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.219409 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a527f913-2977-419c-9a4c-25b4c1729025-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.424819 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-dh7x5"] Nov 22 05:13:08 crc kubenswrapper[4818]: I1122 05:13:08.434843 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-dh7x5"] Nov 22 05:13:09 crc kubenswrapper[4818]: I1122 05:13:09.054373 4818 generic.go:334] "Generic (PLEG): container finished" podID="5a78c342-22c9-41f0-9157-9ca44dcae82f" containerID="125e1c460f3556eca44f4dd4b5178b464cdb2f73230f2728c7faff73312b1ede" exitCode=0 Nov 22 05:13:09 crc kubenswrapper[4818]: I1122 05:13:09.054456 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5a78c342-22c9-41f0-9157-9ca44dcae82f","Type":"ContainerDied","Data":"125e1c460f3556eca44f4dd4b5178b464cdb2f73230f2728c7faff73312b1ede"} Nov 22 05:13:09 crc kubenswrapper[4818]: I1122 05:13:09.058225 4818 generic.go:334] "Generic (PLEG): container finished" podID="10dfb213-2885-405b-8c78-78363efec183" containerID="4f8a0beee11a41b125ccd8bcedeca4bb9b6a58dc64b4e10d3c2f83b5853f9851" exitCode=0 Nov 22 05:13:09 crc kubenswrapper[4818]: I1122 05:13:09.058279 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tfdvb" event={"ID":"10dfb213-2885-405b-8c78-78363efec183","Type":"ContainerDied","Data":"4f8a0beee11a41b125ccd8bcedeca4bb9b6a58dc64b4e10d3c2f83b5853f9851"} Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.012546 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.066034 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5a78c342-22c9-41f0-9157-9ca44dcae82f","Type":"ContainerDied","Data":"d90344a81b4102caa0b1e3859a0722007dacd6c5bce61ebf9548678bff4069bd"} Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.066065 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.066111 4818 scope.go:117] "RemoveContainer" containerID="125e1c460f3556eca44f4dd4b5178b464cdb2f73230f2728c7faff73312b1ede" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.159218 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a78c342-22c9-41f0-9157-9ca44dcae82f-config-data\") pod \"5a78c342-22c9-41f0-9157-9ca44dcae82f\" (UID: \"5a78c342-22c9-41f0-9157-9ca44dcae82f\") " Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.159739 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a78c342-22c9-41f0-9157-9ca44dcae82f-combined-ca-bundle\") pod \"5a78c342-22c9-41f0-9157-9ca44dcae82f\" (UID: \"5a78c342-22c9-41f0-9157-9ca44dcae82f\") " Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.159940 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmgrd\" (UniqueName: \"kubernetes.io/projected/5a78c342-22c9-41f0-9157-9ca44dcae82f-kube-api-access-fmgrd\") pod \"5a78c342-22c9-41f0-9157-9ca44dcae82f\" (UID: \"5a78c342-22c9-41f0-9157-9ca44dcae82f\") " Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.168200 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a78c342-22c9-41f0-9157-9ca44dcae82f-kube-api-access-fmgrd" (OuterVolumeSpecName: "kube-api-access-fmgrd") pod "5a78c342-22c9-41f0-9157-9ca44dcae82f" (UID: "5a78c342-22c9-41f0-9157-9ca44dcae82f"). InnerVolumeSpecName "kube-api-access-fmgrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.189610 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a78c342-22c9-41f0-9157-9ca44dcae82f-config-data" (OuterVolumeSpecName: "config-data") pod "5a78c342-22c9-41f0-9157-9ca44dcae82f" (UID: "5a78c342-22c9-41f0-9157-9ca44dcae82f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.194167 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a78c342-22c9-41f0-9157-9ca44dcae82f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a78c342-22c9-41f0-9157-9ca44dcae82f" (UID: "5a78c342-22c9-41f0-9157-9ca44dcae82f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.262643 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a78c342-22c9-41f0-9157-9ca44dcae82f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.262674 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmgrd\" (UniqueName: \"kubernetes.io/projected/5a78c342-22c9-41f0-9157-9ca44dcae82f-kube-api-access-fmgrd\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.262687 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a78c342-22c9-41f0-9157-9ca44dcae82f-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.302475 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a527f913-2977-419c-9a4c-25b4c1729025" path="/var/lib/kubelet/pods/a527f913-2977-419c-9a4c-25b4c1729025/volumes" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.412276 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tfdvb" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.424106 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.438121 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.461853 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 05:13:10 crc kubenswrapper[4818]: E1122 05:13:10.462751 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a527f913-2977-419c-9a4c-25b4c1729025" containerName="dnsmasq-dns" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.462783 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a527f913-2977-419c-9a4c-25b4c1729025" containerName="dnsmasq-dns" Nov 22 05:13:10 crc kubenswrapper[4818]: E1122 05:13:10.462851 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a527f913-2977-419c-9a4c-25b4c1729025" containerName="init" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.462863 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a527f913-2977-419c-9a4c-25b4c1729025" containerName="init" Nov 22 05:13:10 crc kubenswrapper[4818]: E1122 05:13:10.462917 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7a2b652-b72b-4cd4-85cb-37e996f5238c" containerName="nova-manage" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.462927 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7a2b652-b72b-4cd4-85cb-37e996f5238c" containerName="nova-manage" Nov 22 05:13:10 crc kubenswrapper[4818]: E1122 05:13:10.462944 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10dfb213-2885-405b-8c78-78363efec183" containerName="nova-cell1-conductor-db-sync" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.462952 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="10dfb213-2885-405b-8c78-78363efec183" containerName="nova-cell1-conductor-db-sync" Nov 22 05:13:10 crc kubenswrapper[4818]: E1122 05:13:10.462980 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a78c342-22c9-41f0-9157-9ca44dcae82f" containerName="nova-scheduler-scheduler" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.462989 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a78c342-22c9-41f0-9157-9ca44dcae82f" containerName="nova-scheduler-scheduler" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.463857 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a527f913-2977-419c-9a4c-25b4c1729025" containerName="dnsmasq-dns" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.463884 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a78c342-22c9-41f0-9157-9ca44dcae82f" containerName="nova-scheduler-scheduler" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.463917 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="10dfb213-2885-405b-8c78-78363efec183" containerName="nova-cell1-conductor-db-sync" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.463931 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7a2b652-b72b-4cd4-85cb-37e996f5238c" containerName="nova-manage" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.465087 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.467575 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.473519 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.568684 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-config-data\") pod \"10dfb213-2885-405b-8c78-78363efec183\" (UID: \"10dfb213-2885-405b-8c78-78363efec183\") " Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.568731 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-scripts\") pod \"10dfb213-2885-405b-8c78-78363efec183\" (UID: \"10dfb213-2885-405b-8c78-78363efec183\") " Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.568790 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-combined-ca-bundle\") pod \"10dfb213-2885-405b-8c78-78363efec183\" (UID: \"10dfb213-2885-405b-8c78-78363efec183\") " Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.568825 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gt2w\" (UniqueName: \"kubernetes.io/projected/10dfb213-2885-405b-8c78-78363efec183-kube-api-access-5gt2w\") pod \"10dfb213-2885-405b-8c78-78363efec183\" (UID: \"10dfb213-2885-405b-8c78-78363efec183\") " Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.569079 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bltmn\" (UniqueName: \"kubernetes.io/projected/493288a9-105f-4ac3-8f6e-2fde520d3d19-kube-api-access-bltmn\") pod \"nova-scheduler-0\" (UID: \"493288a9-105f-4ac3-8f6e-2fde520d3d19\") " pod="openstack/nova-scheduler-0" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.569169 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493288a9-105f-4ac3-8f6e-2fde520d3d19-config-data\") pod \"nova-scheduler-0\" (UID: \"493288a9-105f-4ac3-8f6e-2fde520d3d19\") " pod="openstack/nova-scheduler-0" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.569233 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493288a9-105f-4ac3-8f6e-2fde520d3d19-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"493288a9-105f-4ac3-8f6e-2fde520d3d19\") " pod="openstack/nova-scheduler-0" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.573443 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-scripts" (OuterVolumeSpecName: "scripts") pod "10dfb213-2885-405b-8c78-78363efec183" (UID: "10dfb213-2885-405b-8c78-78363efec183"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.573503 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10dfb213-2885-405b-8c78-78363efec183-kube-api-access-5gt2w" (OuterVolumeSpecName: "kube-api-access-5gt2w") pod "10dfb213-2885-405b-8c78-78363efec183" (UID: "10dfb213-2885-405b-8c78-78363efec183"). InnerVolumeSpecName "kube-api-access-5gt2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:13:10 crc kubenswrapper[4818]: E1122 05:13:10.595029 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-combined-ca-bundle podName:10dfb213-2885-405b-8c78-78363efec183 nodeName:}" failed. No retries permitted until 2025-11-22 05:13:11.094959519 +0000 UTC m=+1543.669376046 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-combined-ca-bundle") pod "10dfb213-2885-405b-8c78-78363efec183" (UID: "10dfb213-2885-405b-8c78-78363efec183") : error deleting /var/lib/kubelet/pods/10dfb213-2885-405b-8c78-78363efec183/volume-subpaths: remove /var/lib/kubelet/pods/10dfb213-2885-405b-8c78-78363efec183/volume-subpaths: no such file or directory Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.598708 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-config-data" (OuterVolumeSpecName: "config-data") pod "10dfb213-2885-405b-8c78-78363efec183" (UID: "10dfb213-2885-405b-8c78-78363efec183"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.671230 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493288a9-105f-4ac3-8f6e-2fde520d3d19-config-data\") pod \"nova-scheduler-0\" (UID: \"493288a9-105f-4ac3-8f6e-2fde520d3d19\") " pod="openstack/nova-scheduler-0" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.671362 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493288a9-105f-4ac3-8f6e-2fde520d3d19-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"493288a9-105f-4ac3-8f6e-2fde520d3d19\") " pod="openstack/nova-scheduler-0" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.671456 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bltmn\" (UniqueName: \"kubernetes.io/projected/493288a9-105f-4ac3-8f6e-2fde520d3d19-kube-api-access-bltmn\") pod \"nova-scheduler-0\" (UID: \"493288a9-105f-4ac3-8f6e-2fde520d3d19\") " pod="openstack/nova-scheduler-0" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.672297 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gt2w\" (UniqueName: \"kubernetes.io/projected/10dfb213-2885-405b-8c78-78363efec183-kube-api-access-5gt2w\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.672572 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.672606 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.676154 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493288a9-105f-4ac3-8f6e-2fde520d3d19-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"493288a9-105f-4ac3-8f6e-2fde520d3d19\") " pod="openstack/nova-scheduler-0" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.676302 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493288a9-105f-4ac3-8f6e-2fde520d3d19-config-data\") pod \"nova-scheduler-0\" (UID: \"493288a9-105f-4ac3-8f6e-2fde520d3d19\") " pod="openstack/nova-scheduler-0" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.695596 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bltmn\" (UniqueName: \"kubernetes.io/projected/493288a9-105f-4ac3-8f6e-2fde520d3d19-kube-api-access-bltmn\") pod \"nova-scheduler-0\" (UID: \"493288a9-105f-4ac3-8f6e-2fde520d3d19\") " pod="openstack/nova-scheduler-0" Nov 22 05:13:10 crc kubenswrapper[4818]: I1122 05:13:10.786373 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.082332 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tfdvb" event={"ID":"10dfb213-2885-405b-8c78-78363efec183","Type":"ContainerDied","Data":"dfac1c10ecbf4be3ae77df2eff5811ed0fa0b3dec7527cdffd061d1f32cc1495"} Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.082670 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfac1c10ecbf4be3ae77df2eff5811ed0fa0b3dec7527cdffd061d1f32cc1495" Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.082824 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tfdvb" Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.142207 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.143597 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.155921 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.181846 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-combined-ca-bundle\") pod \"10dfb213-2885-405b-8c78-78363efec183\" (UID: \"10dfb213-2885-405b-8c78-78363efec183\") " Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.182536 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqfvf\" (UniqueName: \"kubernetes.io/projected/f9d0ba71-b442-4c8a-87bd-72a8ca7284b2-kube-api-access-qqfvf\") pod \"nova-cell1-conductor-0\" (UID: \"f9d0ba71-b442-4c8a-87bd-72a8ca7284b2\") " pod="openstack/nova-cell1-conductor-0" Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.182689 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9d0ba71-b442-4c8a-87bd-72a8ca7284b2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f9d0ba71-b442-4c8a-87bd-72a8ca7284b2\") " pod="openstack/nova-cell1-conductor-0" Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.182753 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9d0ba71-b442-4c8a-87bd-72a8ca7284b2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f9d0ba71-b442-4c8a-87bd-72a8ca7284b2\") " pod="openstack/nova-cell1-conductor-0" Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.186390 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "10dfb213-2885-405b-8c78-78363efec183" (UID: "10dfb213-2885-405b-8c78-78363efec183"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.234297 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.284957 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqfvf\" (UniqueName: \"kubernetes.io/projected/f9d0ba71-b442-4c8a-87bd-72a8ca7284b2-kube-api-access-qqfvf\") pod \"nova-cell1-conductor-0\" (UID: \"f9d0ba71-b442-4c8a-87bd-72a8ca7284b2\") " pod="openstack/nova-cell1-conductor-0" Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.285760 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9d0ba71-b442-4c8a-87bd-72a8ca7284b2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f9d0ba71-b442-4c8a-87bd-72a8ca7284b2\") " pod="openstack/nova-cell1-conductor-0" Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.285804 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9d0ba71-b442-4c8a-87bd-72a8ca7284b2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f9d0ba71-b442-4c8a-87bd-72a8ca7284b2\") " pod="openstack/nova-cell1-conductor-0" Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.285944 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10dfb213-2885-405b-8c78-78363efec183-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.291466 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9d0ba71-b442-4c8a-87bd-72a8ca7284b2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f9d0ba71-b442-4c8a-87bd-72a8ca7284b2\") " pod="openstack/nova-cell1-conductor-0" Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.296130 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9d0ba71-b442-4c8a-87bd-72a8ca7284b2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f9d0ba71-b442-4c8a-87bd-72a8ca7284b2\") " pod="openstack/nova-cell1-conductor-0" Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.303856 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqfvf\" (UniqueName: \"kubernetes.io/projected/f9d0ba71-b442-4c8a-87bd-72a8ca7284b2-kube-api-access-qqfvf\") pod \"nova-cell1-conductor-0\" (UID: \"f9d0ba71-b442-4c8a-87bd-72a8ca7284b2\") " pod="openstack/nova-cell1-conductor-0" Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.473441 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 22 05:13:11 crc kubenswrapper[4818]: I1122 05:13:11.903363 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 22 05:13:12 crc kubenswrapper[4818]: I1122 05:13:12.091185 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"493288a9-105f-4ac3-8f6e-2fde520d3d19","Type":"ContainerStarted","Data":"4f6b54fda6fd2e611e1267b94b1d33df9e78737c66fdc769a1fb78e46a2fc845"} Nov 22 05:13:12 crc kubenswrapper[4818]: I1122 05:13:12.092392 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"493288a9-105f-4ac3-8f6e-2fde520d3d19","Type":"ContainerStarted","Data":"2d5ce36095c8d050046b26fd9171d0df09f2ad5e5da1bb6dc5b07dbf4b957fca"} Nov 22 05:13:12 crc kubenswrapper[4818]: I1122 05:13:12.096709 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"f9d0ba71-b442-4c8a-87bd-72a8ca7284b2","Type":"ContainerStarted","Data":"758690a65ae3e0cd69591764bbb25e59444b0fb060239391d897ce803f40181d"} Nov 22 05:13:12 crc kubenswrapper[4818]: I1122 05:13:12.114457 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.114432304 podStartE2EDuration="2.114432304s" podCreationTimestamp="2025-11-22 05:13:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:13:12.105373544 +0000 UTC m=+1544.679790091" watchObservedRunningTime="2025-11-22 05:13:12.114432304 +0000 UTC m=+1544.688848841" Nov 22 05:13:12 crc kubenswrapper[4818]: I1122 05:13:12.303746 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a78c342-22c9-41f0-9157-9ca44dcae82f" path="/var/lib/kubelet/pods/5a78c342-22c9-41f0-9157-9ca44dcae82f/volumes" Nov 22 05:13:13 crc kubenswrapper[4818]: I1122 05:13:13.116447 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"f9d0ba71-b442-4c8a-87bd-72a8ca7284b2","Type":"ContainerStarted","Data":"1f76a3eccd9b591e41a19e8910c3ac533ed14f3f1d43562f0dd7ad1a04656a47"} Nov 22 05:13:13 crc kubenswrapper[4818]: I1122 05:13:13.117333 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 22 05:13:13 crc kubenswrapper[4818]: I1122 05:13:13.143468 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.143450527 podStartE2EDuration="2.143450527s" podCreationTimestamp="2025-11-22 05:13:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:13:13.138038034 +0000 UTC m=+1545.712454651" watchObservedRunningTime="2025-11-22 05:13:13.143450527 +0000 UTC m=+1545.717867054" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.007421 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.049074 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3df5198-1e08-4baf-811f-493f782a8ed1-logs\") pod \"b3df5198-1e08-4baf-811f-493f782a8ed1\" (UID: \"b3df5198-1e08-4baf-811f-493f782a8ed1\") " Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.049149 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3df5198-1e08-4baf-811f-493f782a8ed1-config-data\") pod \"b3df5198-1e08-4baf-811f-493f782a8ed1\" (UID: \"b3df5198-1e08-4baf-811f-493f782a8ed1\") " Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.049181 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b95kf\" (UniqueName: \"kubernetes.io/projected/b3df5198-1e08-4baf-811f-493f782a8ed1-kube-api-access-b95kf\") pod \"b3df5198-1e08-4baf-811f-493f782a8ed1\" (UID: \"b3df5198-1e08-4baf-811f-493f782a8ed1\") " Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.049228 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3df5198-1e08-4baf-811f-493f782a8ed1-combined-ca-bundle\") pod \"b3df5198-1e08-4baf-811f-493f782a8ed1\" (UID: \"b3df5198-1e08-4baf-811f-493f782a8ed1\") " Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.049669 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3df5198-1e08-4baf-811f-493f782a8ed1-logs" (OuterVolumeSpecName: "logs") pod "b3df5198-1e08-4baf-811f-493f782a8ed1" (UID: "b3df5198-1e08-4baf-811f-493f782a8ed1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.056871 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3df5198-1e08-4baf-811f-493f782a8ed1-kube-api-access-b95kf" (OuterVolumeSpecName: "kube-api-access-b95kf") pod "b3df5198-1e08-4baf-811f-493f782a8ed1" (UID: "b3df5198-1e08-4baf-811f-493f782a8ed1"). InnerVolumeSpecName "kube-api-access-b95kf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.084404 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3df5198-1e08-4baf-811f-493f782a8ed1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3df5198-1e08-4baf-811f-493f782a8ed1" (UID: "b3df5198-1e08-4baf-811f-493f782a8ed1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.098333 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3df5198-1e08-4baf-811f-493f782a8ed1-config-data" (OuterVolumeSpecName: "config-data") pod "b3df5198-1e08-4baf-811f-493f782a8ed1" (UID: "b3df5198-1e08-4baf-811f-493f782a8ed1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.126623 4818 generic.go:334] "Generic (PLEG): container finished" podID="b3df5198-1e08-4baf-811f-493f782a8ed1" containerID="9fc067ba8f3205ac4e510123b4beddd9eff6a6fc1b1ffb3cd4d6ec2d9e204d63" exitCode=0 Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.126664 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b3df5198-1e08-4baf-811f-493f782a8ed1","Type":"ContainerDied","Data":"9fc067ba8f3205ac4e510123b4beddd9eff6a6fc1b1ffb3cd4d6ec2d9e204d63"} Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.126712 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.126734 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b3df5198-1e08-4baf-811f-493f782a8ed1","Type":"ContainerDied","Data":"db0e78bb921c92dc07d69e0574a9c0d9d9e0a7dc88fe3f8148c2a02ccc2c1e87"} Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.126762 4818 scope.go:117] "RemoveContainer" containerID="9fc067ba8f3205ac4e510123b4beddd9eff6a6fc1b1ffb3cd4d6ec2d9e204d63" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.153496 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.156066 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3df5198-1e08-4baf-811f-493f782a8ed1-logs\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.156098 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3df5198-1e08-4baf-811f-493f782a8ed1-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.156109 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b95kf\" (UniqueName: \"kubernetes.io/projected/b3df5198-1e08-4baf-811f-493f782a8ed1-kube-api-access-b95kf\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.156122 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3df5198-1e08-4baf-811f-493f782a8ed1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.156610 4818 scope.go:117] "RemoveContainer" containerID="3b9d7e170741eedb47592033e9a1d7b80e86fb1648481543e959d67c29bd8084" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.167992 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.177881 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 22 05:13:14 crc kubenswrapper[4818]: E1122 05:13:14.178477 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3df5198-1e08-4baf-811f-493f782a8ed1" containerName="nova-api-log" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.178496 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3df5198-1e08-4baf-811f-493f782a8ed1" containerName="nova-api-log" Nov 22 05:13:14 crc kubenswrapper[4818]: E1122 05:13:14.178526 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3df5198-1e08-4baf-811f-493f782a8ed1" containerName="nova-api-api" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.178533 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3df5198-1e08-4baf-811f-493f782a8ed1" containerName="nova-api-api" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.178691 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3df5198-1e08-4baf-811f-493f782a8ed1" containerName="nova-api-log" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.185663 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3df5198-1e08-4baf-811f-493f782a8ed1" containerName="nova-api-api" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.187072 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.189210 4818 scope.go:117] "RemoveContainer" containerID="9fc067ba8f3205ac4e510123b4beddd9eff6a6fc1b1ffb3cd4d6ec2d9e204d63" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.191783 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.198016 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 05:13:14 crc kubenswrapper[4818]: E1122 05:13:14.247656 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fc067ba8f3205ac4e510123b4beddd9eff6a6fc1b1ffb3cd4d6ec2d9e204d63\": container with ID starting with 9fc067ba8f3205ac4e510123b4beddd9eff6a6fc1b1ffb3cd4d6ec2d9e204d63 not found: ID does not exist" containerID="9fc067ba8f3205ac4e510123b4beddd9eff6a6fc1b1ffb3cd4d6ec2d9e204d63" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.247731 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fc067ba8f3205ac4e510123b4beddd9eff6a6fc1b1ffb3cd4d6ec2d9e204d63"} err="failed to get container status \"9fc067ba8f3205ac4e510123b4beddd9eff6a6fc1b1ffb3cd4d6ec2d9e204d63\": rpc error: code = NotFound desc = could not find container \"9fc067ba8f3205ac4e510123b4beddd9eff6a6fc1b1ffb3cd4d6ec2d9e204d63\": container with ID starting with 9fc067ba8f3205ac4e510123b4beddd9eff6a6fc1b1ffb3cd4d6ec2d9e204d63 not found: ID does not exist" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.247761 4818 scope.go:117] "RemoveContainer" containerID="3b9d7e170741eedb47592033e9a1d7b80e86fb1648481543e959d67c29bd8084" Nov 22 05:13:14 crc kubenswrapper[4818]: E1122 05:13:14.248316 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b9d7e170741eedb47592033e9a1d7b80e86fb1648481543e959d67c29bd8084\": container with ID starting with 3b9d7e170741eedb47592033e9a1d7b80e86fb1648481543e959d67c29bd8084 not found: ID does not exist" containerID="3b9d7e170741eedb47592033e9a1d7b80e86fb1648481543e959d67c29bd8084" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.248361 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b9d7e170741eedb47592033e9a1d7b80e86fb1648481543e959d67c29bd8084"} err="failed to get container status \"3b9d7e170741eedb47592033e9a1d7b80e86fb1648481543e959d67c29bd8084\": rpc error: code = NotFound desc = could not find container \"3b9d7e170741eedb47592033e9a1d7b80e86fb1648481543e959d67c29bd8084\": container with ID starting with 3b9d7e170741eedb47592033e9a1d7b80e86fb1648481543e959d67c29bd8084 not found: ID does not exist" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.257855 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmnwr\" (UniqueName: \"kubernetes.io/projected/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-kube-api-access-tmnwr\") pod \"nova-api-0\" (UID: \"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0\") " pod="openstack/nova-api-0" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.257960 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0\") " pod="openstack/nova-api-0" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.257988 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-logs\") pod \"nova-api-0\" (UID: \"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0\") " pod="openstack/nova-api-0" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.258034 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-config-data\") pod \"nova-api-0\" (UID: \"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0\") " pod="openstack/nova-api-0" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.305640 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3df5198-1e08-4baf-811f-493f782a8ed1" path="/var/lib/kubelet/pods/b3df5198-1e08-4baf-811f-493f782a8ed1/volumes" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.359258 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0\") " pod="openstack/nova-api-0" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.359354 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-logs\") pod \"nova-api-0\" (UID: \"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0\") " pod="openstack/nova-api-0" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.359406 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-config-data\") pod \"nova-api-0\" (UID: \"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0\") " pod="openstack/nova-api-0" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.359461 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmnwr\" (UniqueName: \"kubernetes.io/projected/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-kube-api-access-tmnwr\") pod \"nova-api-0\" (UID: \"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0\") " pod="openstack/nova-api-0" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.359996 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-logs\") pod \"nova-api-0\" (UID: \"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0\") " pod="openstack/nova-api-0" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.365088 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0\") " pod="openstack/nova-api-0" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.365739 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-config-data\") pod \"nova-api-0\" (UID: \"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0\") " pod="openstack/nova-api-0" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.377889 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmnwr\" (UniqueName: \"kubernetes.io/projected/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-kube-api-access-tmnwr\") pod \"nova-api-0\" (UID: \"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0\") " pod="openstack/nova-api-0" Nov 22 05:13:14 crc kubenswrapper[4818]: I1122 05:13:14.568917 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 05:13:15 crc kubenswrapper[4818]: I1122 05:13:15.028727 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 05:13:15 crc kubenswrapper[4818]: W1122 05:13:15.030539 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5998a0df_9ea3_4d81_92d8_01f4c9d7e2e0.slice/crio-46eb631546a92c46c242a7ed047cb5097613760c2ade558bf0acc9a57f584cfc WatchSource:0}: Error finding container 46eb631546a92c46c242a7ed047cb5097613760c2ade558bf0acc9a57f584cfc: Status 404 returned error can't find the container with id 46eb631546a92c46c242a7ed047cb5097613760c2ade558bf0acc9a57f584cfc Nov 22 05:13:15 crc kubenswrapper[4818]: I1122 05:13:15.137127 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0","Type":"ContainerStarted","Data":"46eb631546a92c46c242a7ed047cb5097613760c2ade558bf0acc9a57f584cfc"} Nov 22 05:13:15 crc kubenswrapper[4818]: I1122 05:13:15.786799 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 22 05:13:16 crc kubenswrapper[4818]: I1122 05:13:16.157018 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0","Type":"ContainerStarted","Data":"fc69b2d8313b58d3dee5a61cbf5df5773080dd416463c96453502e8479bee49c"} Nov 22 05:13:16 crc kubenswrapper[4818]: I1122 05:13:16.157401 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0","Type":"ContainerStarted","Data":"d20be45b2a4e0531fc49afd0ee6f16b2e30c5f3ab71110e1a37d338aedbdac59"} Nov 22 05:13:16 crc kubenswrapper[4818]: I1122 05:13:16.190855 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.19083655 podStartE2EDuration="2.19083655s" podCreationTimestamp="2025-11-22 05:13:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:13:16.178657669 +0000 UTC m=+1548.753074216" watchObservedRunningTime="2025-11-22 05:13:16.19083655 +0000 UTC m=+1548.765253077" Nov 22 05:13:20 crc kubenswrapper[4818]: I1122 05:13:20.787070 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 22 05:13:20 crc kubenswrapper[4818]: I1122 05:13:20.832606 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 22 05:13:21 crc kubenswrapper[4818]: I1122 05:13:21.265079 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 22 05:13:21 crc kubenswrapper[4818]: I1122 05:13:21.519772 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 22 05:13:24 crc kubenswrapper[4818]: I1122 05:13:24.570239 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 05:13:24 crc kubenswrapper[4818]: I1122 05:13:24.570811 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 05:13:25 crc kubenswrapper[4818]: I1122 05:13:25.654485 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.177:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 05:13:25 crc kubenswrapper[4818]: I1122 05:13:25.654508 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.177:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.351486 4818 generic.go:334] "Generic (PLEG): container finished" podID="e78732aa-3735-4262-bf33-2726fd0a6ff1" containerID="7f2b20c7ad72020b50ab18a72c91647f36d316ff9d9b4007722aabac0d1e6b9d" exitCode=137 Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.351976 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e78732aa-3735-4262-bf33-2726fd0a6ff1","Type":"ContainerDied","Data":"7f2b20c7ad72020b50ab18a72c91647f36d316ff9d9b4007722aabac0d1e6b9d"} Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.353647 4818 generic.go:334] "Generic (PLEG): container finished" podID="cb1ca6ba-4bdf-44d4-8478-59d41abec5d7" containerID="7f6074bbb86e476297d4ead08a31f1c8c1400eed778d33e5b8b10eb527e58d43" exitCode=137 Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.353682 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cb1ca6ba-4bdf-44d4-8478-59d41abec5d7","Type":"ContainerDied","Data":"7f6074bbb86e476297d4ead08a31f1c8c1400eed778d33e5b8b10eb527e58d43"} Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.477413 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.481798 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.598133 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e78732aa-3735-4262-bf33-2726fd0a6ff1-logs\") pod \"e78732aa-3735-4262-bf33-2726fd0a6ff1\" (UID: \"e78732aa-3735-4262-bf33-2726fd0a6ff1\") " Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.598314 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncx9g\" (UniqueName: \"kubernetes.io/projected/e78732aa-3735-4262-bf33-2726fd0a6ff1-kube-api-access-ncx9g\") pod \"e78732aa-3735-4262-bf33-2726fd0a6ff1\" (UID: \"e78732aa-3735-4262-bf33-2726fd0a6ff1\") " Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.598355 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbpnh\" (UniqueName: \"kubernetes.io/projected/cb1ca6ba-4bdf-44d4-8478-59d41abec5d7-kube-api-access-qbpnh\") pod \"cb1ca6ba-4bdf-44d4-8478-59d41abec5d7\" (UID: \"cb1ca6ba-4bdf-44d4-8478-59d41abec5d7\") " Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.598515 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e78732aa-3735-4262-bf33-2726fd0a6ff1-config-data\") pod \"e78732aa-3735-4262-bf33-2726fd0a6ff1\" (UID: \"e78732aa-3735-4262-bf33-2726fd0a6ff1\") " Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.598599 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb1ca6ba-4bdf-44d4-8478-59d41abec5d7-config-data\") pod \"cb1ca6ba-4bdf-44d4-8478-59d41abec5d7\" (UID: \"cb1ca6ba-4bdf-44d4-8478-59d41abec5d7\") " Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.598654 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb1ca6ba-4bdf-44d4-8478-59d41abec5d7-combined-ca-bundle\") pod \"cb1ca6ba-4bdf-44d4-8478-59d41abec5d7\" (UID: \"cb1ca6ba-4bdf-44d4-8478-59d41abec5d7\") " Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.598681 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e78732aa-3735-4262-bf33-2726fd0a6ff1-combined-ca-bundle\") pod \"e78732aa-3735-4262-bf33-2726fd0a6ff1\" (UID: \"e78732aa-3735-4262-bf33-2726fd0a6ff1\") " Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.598807 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e78732aa-3735-4262-bf33-2726fd0a6ff1-logs" (OuterVolumeSpecName: "logs") pod "e78732aa-3735-4262-bf33-2726fd0a6ff1" (UID: "e78732aa-3735-4262-bf33-2726fd0a6ff1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.599126 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e78732aa-3735-4262-bf33-2726fd0a6ff1-logs\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.604519 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e78732aa-3735-4262-bf33-2726fd0a6ff1-kube-api-access-ncx9g" (OuterVolumeSpecName: "kube-api-access-ncx9g") pod "e78732aa-3735-4262-bf33-2726fd0a6ff1" (UID: "e78732aa-3735-4262-bf33-2726fd0a6ff1"). InnerVolumeSpecName "kube-api-access-ncx9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.604575 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb1ca6ba-4bdf-44d4-8478-59d41abec5d7-kube-api-access-qbpnh" (OuterVolumeSpecName: "kube-api-access-qbpnh") pod "cb1ca6ba-4bdf-44d4-8478-59d41abec5d7" (UID: "cb1ca6ba-4bdf-44d4-8478-59d41abec5d7"). InnerVolumeSpecName "kube-api-access-qbpnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.627106 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb1ca6ba-4bdf-44d4-8478-59d41abec5d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb1ca6ba-4bdf-44d4-8478-59d41abec5d7" (UID: "cb1ca6ba-4bdf-44d4-8478-59d41abec5d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.629802 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb1ca6ba-4bdf-44d4-8478-59d41abec5d7-config-data" (OuterVolumeSpecName: "config-data") pod "cb1ca6ba-4bdf-44d4-8478-59d41abec5d7" (UID: "cb1ca6ba-4bdf-44d4-8478-59d41abec5d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.630421 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e78732aa-3735-4262-bf33-2726fd0a6ff1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e78732aa-3735-4262-bf33-2726fd0a6ff1" (UID: "e78732aa-3735-4262-bf33-2726fd0a6ff1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.632795 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e78732aa-3735-4262-bf33-2726fd0a6ff1-config-data" (OuterVolumeSpecName: "config-data") pod "e78732aa-3735-4262-bf33-2726fd0a6ff1" (UID: "e78732aa-3735-4262-bf33-2726fd0a6ff1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.701469 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb1ca6ba-4bdf-44d4-8478-59d41abec5d7-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.701524 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb1ca6ba-4bdf-44d4-8478-59d41abec5d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.701546 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e78732aa-3735-4262-bf33-2726fd0a6ff1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.701564 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncx9g\" (UniqueName: \"kubernetes.io/projected/e78732aa-3735-4262-bf33-2726fd0a6ff1-kube-api-access-ncx9g\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.701582 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbpnh\" (UniqueName: \"kubernetes.io/projected/cb1ca6ba-4bdf-44d4-8478-59d41abec5d7-kube-api-access-qbpnh\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:32 crc kubenswrapper[4818]: I1122 05:13:32.701601 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e78732aa-3735-4262-bf33-2726fd0a6ff1-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.366867 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e78732aa-3735-4262-bf33-2726fd0a6ff1","Type":"ContainerDied","Data":"057e8530e64d5501a445e4514710c54a3028346b64cebf7f2245814b4a8725d6"} Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.366914 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.366935 4818 scope.go:117] "RemoveContainer" containerID="7f2b20c7ad72020b50ab18a72c91647f36d316ff9d9b4007722aabac0d1e6b9d" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.370327 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.370837 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cb1ca6ba-4bdf-44d4-8478-59d41abec5d7","Type":"ContainerDied","Data":"08401c280cc3e65837f99f61db45e33c90372a256a33426ef9ecdb0fa16a4dd7"} Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.394362 4818 scope.go:117] "RemoveContainer" containerID="17684c1a5c2c94a609e2c8edff7f798e0d6a65ca7f711fdf7ce896fba696da5e" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.415732 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.427878 4818 scope.go:117] "RemoveContainer" containerID="7f6074bbb86e476297d4ead08a31f1c8c1400eed778d33e5b8b10eb527e58d43" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.453415 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.463660 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.472226 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.480097 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 05:13:33 crc kubenswrapper[4818]: E1122 05:13:33.480555 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78732aa-3735-4262-bf33-2726fd0a6ff1" containerName="nova-metadata-log" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.480577 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78732aa-3735-4262-bf33-2726fd0a6ff1" containerName="nova-metadata-log" Nov 22 05:13:33 crc kubenswrapper[4818]: E1122 05:13:33.480600 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb1ca6ba-4bdf-44d4-8478-59d41abec5d7" containerName="nova-cell1-novncproxy-novncproxy" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.480607 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb1ca6ba-4bdf-44d4-8478-59d41abec5d7" containerName="nova-cell1-novncproxy-novncproxy" Nov 22 05:13:33 crc kubenswrapper[4818]: E1122 05:13:33.480620 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78732aa-3735-4262-bf33-2726fd0a6ff1" containerName="nova-metadata-metadata" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.480626 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78732aa-3735-4262-bf33-2726fd0a6ff1" containerName="nova-metadata-metadata" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.480792 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb1ca6ba-4bdf-44d4-8478-59d41abec5d7" containerName="nova-cell1-novncproxy-novncproxy" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.480809 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="e78732aa-3735-4262-bf33-2726fd0a6ff1" containerName="nova-metadata-log" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.480825 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="e78732aa-3735-4262-bf33-2726fd0a6ff1" containerName="nova-metadata-metadata" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.481507 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.487653 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.489464 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.489815 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.489972 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.490230 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.491647 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.493750 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.496010 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.505899 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.621366 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggwtr\" (UniqueName: \"kubernetes.io/projected/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-kube-api-access-ggwtr\") pod \"nova-metadata-0\" (UID: \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\") " pod="openstack/nova-metadata-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.621473 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-logs\") pod \"nova-metadata-0\" (UID: \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\") " pod="openstack/nova-metadata-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.621551 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f5dcaf3-117e-4024-b1fc-41878a0eefd8-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f5dcaf3-117e-4024-b1fc-41878a0eefd8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.621620 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\") " pod="openstack/nova-metadata-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.621722 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f5dcaf3-117e-4024-b1fc-41878a0eefd8-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f5dcaf3-117e-4024-b1fc-41878a0eefd8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.621783 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\") " pod="openstack/nova-metadata-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.621867 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-config-data\") pod \"nova-metadata-0\" (UID: \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\") " pod="openstack/nova-metadata-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.621975 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2x5v\" (UniqueName: \"kubernetes.io/projected/3f5dcaf3-117e-4024-b1fc-41878a0eefd8-kube-api-access-c2x5v\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f5dcaf3-117e-4024-b1fc-41878a0eefd8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.622007 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f5dcaf3-117e-4024-b1fc-41878a0eefd8-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f5dcaf3-117e-4024-b1fc-41878a0eefd8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.622037 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f5dcaf3-117e-4024-b1fc-41878a0eefd8-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f5dcaf3-117e-4024-b1fc-41878a0eefd8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.724911 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggwtr\" (UniqueName: \"kubernetes.io/projected/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-kube-api-access-ggwtr\") pod \"nova-metadata-0\" (UID: \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\") " pod="openstack/nova-metadata-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.725103 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-logs\") pod \"nova-metadata-0\" (UID: \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\") " pod="openstack/nova-metadata-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.725212 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f5dcaf3-117e-4024-b1fc-41878a0eefd8-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f5dcaf3-117e-4024-b1fc-41878a0eefd8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.725312 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\") " pod="openstack/nova-metadata-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.725483 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f5dcaf3-117e-4024-b1fc-41878a0eefd8-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f5dcaf3-117e-4024-b1fc-41878a0eefd8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.725580 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\") " pod="openstack/nova-metadata-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.727016 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-logs\") pod \"nova-metadata-0\" (UID: \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\") " pod="openstack/nova-metadata-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.727539 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-config-data\") pod \"nova-metadata-0\" (UID: \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\") " pod="openstack/nova-metadata-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.727941 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2x5v\" (UniqueName: \"kubernetes.io/projected/3f5dcaf3-117e-4024-b1fc-41878a0eefd8-kube-api-access-c2x5v\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f5dcaf3-117e-4024-b1fc-41878a0eefd8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.728051 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f5dcaf3-117e-4024-b1fc-41878a0eefd8-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f5dcaf3-117e-4024-b1fc-41878a0eefd8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.728175 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f5dcaf3-117e-4024-b1fc-41878a0eefd8-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f5dcaf3-117e-4024-b1fc-41878a0eefd8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.731095 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\") " pod="openstack/nova-metadata-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.731818 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\") " pod="openstack/nova-metadata-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.732052 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f5dcaf3-117e-4024-b1fc-41878a0eefd8-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f5dcaf3-117e-4024-b1fc-41878a0eefd8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.732684 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-config-data\") pod \"nova-metadata-0\" (UID: \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\") " pod="openstack/nova-metadata-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.732902 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f5dcaf3-117e-4024-b1fc-41878a0eefd8-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f5dcaf3-117e-4024-b1fc-41878a0eefd8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.734062 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f5dcaf3-117e-4024-b1fc-41878a0eefd8-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f5dcaf3-117e-4024-b1fc-41878a0eefd8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.740429 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f5dcaf3-117e-4024-b1fc-41878a0eefd8-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f5dcaf3-117e-4024-b1fc-41878a0eefd8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.740584 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggwtr\" (UniqueName: \"kubernetes.io/projected/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-kube-api-access-ggwtr\") pod \"nova-metadata-0\" (UID: \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\") " pod="openstack/nova-metadata-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.745190 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2x5v\" (UniqueName: \"kubernetes.io/projected/3f5dcaf3-117e-4024-b1fc-41878a0eefd8-kube-api-access-c2x5v\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f5dcaf3-117e-4024-b1fc-41878a0eefd8\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.810511 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:33 crc kubenswrapper[4818]: I1122 05:13:33.819523 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 05:13:34 crc kubenswrapper[4818]: I1122 05:13:34.321309 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb1ca6ba-4bdf-44d4-8478-59d41abec5d7" path="/var/lib/kubelet/pods/cb1ca6ba-4bdf-44d4-8478-59d41abec5d7/volumes" Nov 22 05:13:34 crc kubenswrapper[4818]: I1122 05:13:34.323995 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e78732aa-3735-4262-bf33-2726fd0a6ff1" path="/var/lib/kubelet/pods/e78732aa-3735-4262-bf33-2726fd0a6ff1/volumes" Nov 22 05:13:34 crc kubenswrapper[4818]: I1122 05:13:34.325065 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 05:13:34 crc kubenswrapper[4818]: I1122 05:13:34.328966 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 05:13:34 crc kubenswrapper[4818]: W1122 05:13:34.330438 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a6786f7_e9e0_4bcc_9115_1bfbbf7e8b12.slice/crio-ee952d2e76d4047b2d0404e9f0bc056a3707dc8795bbe5ad8b5d3876a0d4e6dd WatchSource:0}: Error finding container ee952d2e76d4047b2d0404e9f0bc056a3707dc8795bbe5ad8b5d3876a0d4e6dd: Status 404 returned error can't find the container with id ee952d2e76d4047b2d0404e9f0bc056a3707dc8795bbe5ad8b5d3876a0d4e6dd Nov 22 05:13:34 crc kubenswrapper[4818]: I1122 05:13:34.386040 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3f5dcaf3-117e-4024-b1fc-41878a0eefd8","Type":"ContainerStarted","Data":"95982152e1c321e5160d69003597a35d523e8055f6d01cf5402b2eb7b7952dac"} Nov 22 05:13:34 crc kubenswrapper[4818]: I1122 05:13:34.388190 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12","Type":"ContainerStarted","Data":"ee952d2e76d4047b2d0404e9f0bc056a3707dc8795bbe5ad8b5d3876a0d4e6dd"} Nov 22 05:13:34 crc kubenswrapper[4818]: I1122 05:13:34.574760 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 22 05:13:34 crc kubenswrapper[4818]: I1122 05:13:34.575479 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 22 05:13:34 crc kubenswrapper[4818]: I1122 05:13:34.575598 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 22 05:13:34 crc kubenswrapper[4818]: I1122 05:13:34.583852 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.396860 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3f5dcaf3-117e-4024-b1fc-41878a0eefd8","Type":"ContainerStarted","Data":"de43be855ee6cb31c30420a5e8c4a6c3a3be2cfc1c5e09930d59fc935732f2e5"} Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.401755 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12","Type":"ContainerStarted","Data":"51dd0c67d382bf1b8bbe35e6cac4fe11396139cdb92c92f82e2aeab8dcf57231"} Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.401784 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12","Type":"ContainerStarted","Data":"291f8def5fd201497ecca5e29c50e6eed2daff4a131454883b80d38d73a92927"} Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.402290 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.416158 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.421573 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.421552573 podStartE2EDuration="2.421552573s" podCreationTimestamp="2025-11-22 05:13:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:13:35.416450527 +0000 UTC m=+1567.990867054" watchObservedRunningTime="2025-11-22 05:13:35.421552573 +0000 UTC m=+1567.995969100" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.465331 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.465304683 podStartE2EDuration="2.465304683s" podCreationTimestamp="2025-11-22 05:13:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:13:35.450378558 +0000 UTC m=+1568.024795115" watchObservedRunningTime="2025-11-22 05:13:35.465304683 +0000 UTC m=+1568.039721220" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.596385 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-88nx2"] Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.597957 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.629689 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-88nx2"] Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.678047 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-ovsdbserver-sb\") pod \"dnsmasq-dns-5b856c5697-88nx2\" (UID: \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\") " pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.678154 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-config\") pod \"dnsmasq-dns-5b856c5697-88nx2\" (UID: \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\") " pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.678187 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-dns-svc\") pod \"dnsmasq-dns-5b856c5697-88nx2\" (UID: \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\") " pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.678202 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qz776\" (UniqueName: \"kubernetes.io/projected/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-kube-api-access-qz776\") pod \"dnsmasq-dns-5b856c5697-88nx2\" (UID: \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\") " pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.678283 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-ovsdbserver-nb\") pod \"dnsmasq-dns-5b856c5697-88nx2\" (UID: \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\") " pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.780504 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-config\") pod \"dnsmasq-dns-5b856c5697-88nx2\" (UID: \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\") " pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.780928 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-dns-svc\") pod \"dnsmasq-dns-5b856c5697-88nx2\" (UID: \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\") " pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.780954 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz776\" (UniqueName: \"kubernetes.io/projected/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-kube-api-access-qz776\") pod \"dnsmasq-dns-5b856c5697-88nx2\" (UID: \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\") " pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.780992 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-ovsdbserver-nb\") pod \"dnsmasq-dns-5b856c5697-88nx2\" (UID: \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\") " pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.781074 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-ovsdbserver-sb\") pod \"dnsmasq-dns-5b856c5697-88nx2\" (UID: \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\") " pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.781749 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-config\") pod \"dnsmasq-dns-5b856c5697-88nx2\" (UID: \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\") " pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.781971 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-dns-svc\") pod \"dnsmasq-dns-5b856c5697-88nx2\" (UID: \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\") " pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.782124 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-ovsdbserver-sb\") pod \"dnsmasq-dns-5b856c5697-88nx2\" (UID: \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\") " pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.782213 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-ovsdbserver-nb\") pod \"dnsmasq-dns-5b856c5697-88nx2\" (UID: \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\") " pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.812173 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qz776\" (UniqueName: \"kubernetes.io/projected/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-kube-api-access-qz776\") pod \"dnsmasq-dns-5b856c5697-88nx2\" (UID: \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\") " pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:13:35 crc kubenswrapper[4818]: I1122 05:13:35.930345 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:13:36 crc kubenswrapper[4818]: I1122 05:13:36.387544 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-88nx2"] Nov 22 05:13:36 crc kubenswrapper[4818]: W1122 05:13:36.395568 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ca5e50d_5880_40a9_a08b_d8c22ad64e0b.slice/crio-702fcd598776c9708585339c988b081837b060ec6085bd9ff4b8919251ceaca0 WatchSource:0}: Error finding container 702fcd598776c9708585339c988b081837b060ec6085bd9ff4b8919251ceaca0: Status 404 returned error can't find the container with id 702fcd598776c9708585339c988b081837b060ec6085bd9ff4b8919251ceaca0 Nov 22 05:13:36 crc kubenswrapper[4818]: I1122 05:13:36.411704 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-88nx2" event={"ID":"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b","Type":"ContainerStarted","Data":"702fcd598776c9708585339c988b081837b060ec6085bd9ff4b8919251ceaca0"} Nov 22 05:13:37 crc kubenswrapper[4818]: I1122 05:13:37.421857 4818 generic.go:334] "Generic (PLEG): container finished" podID="4ca5e50d-5880-40a9-a08b-d8c22ad64e0b" containerID="65a098c1be181f5c1aa80954953218c68c73bcd173598839fc40c230b11e130d" exitCode=0 Nov 22 05:13:37 crc kubenswrapper[4818]: I1122 05:13:37.421920 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-88nx2" event={"ID":"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b","Type":"ContainerDied","Data":"65a098c1be181f5c1aa80954953218c68c73bcd173598839fc40c230b11e130d"} Nov 22 05:13:37 crc kubenswrapper[4818]: I1122 05:13:37.668848 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:13:37 crc kubenswrapper[4818]: I1122 05:13:37.669395 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="501083a5-d8f1-4171-8eca-c331c883ddb9" containerName="ceilometer-central-agent" containerID="cri-o://35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c" gracePeriod=30 Nov 22 05:13:37 crc kubenswrapper[4818]: I1122 05:13:37.669539 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="501083a5-d8f1-4171-8eca-c331c883ddb9" containerName="ceilometer-notification-agent" containerID="cri-o://a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6" gracePeriod=30 Nov 22 05:13:37 crc kubenswrapper[4818]: I1122 05:13:37.669571 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="501083a5-d8f1-4171-8eca-c331c883ddb9" containerName="proxy-httpd" containerID="cri-o://8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344" gracePeriod=30 Nov 22 05:13:37 crc kubenswrapper[4818]: I1122 05:13:37.669544 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="501083a5-d8f1-4171-8eca-c331c883ddb9" containerName="sg-core" containerID="cri-o://3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2" gracePeriod=30 Nov 22 05:13:37 crc kubenswrapper[4818]: I1122 05:13:37.811405 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.356815 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.441314 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-88nx2" event={"ID":"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b","Type":"ContainerStarted","Data":"6de19732be39a975d665120b9409240a0ac7f79dee458e2e995c4b2ae325abc5"} Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.442558 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.449063 4818 generic.go:334] "Generic (PLEG): container finished" podID="501083a5-d8f1-4171-8eca-c331c883ddb9" containerID="8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344" exitCode=0 Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.449104 4818 generic.go:334] "Generic (PLEG): container finished" podID="501083a5-d8f1-4171-8eca-c331c883ddb9" containerID="3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2" exitCode=2 Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.449111 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.449116 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"501083a5-d8f1-4171-8eca-c331c883ddb9","Type":"ContainerDied","Data":"8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344"} Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.449153 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"501083a5-d8f1-4171-8eca-c331c883ddb9","Type":"ContainerDied","Data":"3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2"} Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.449176 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"501083a5-d8f1-4171-8eca-c331c883ddb9","Type":"ContainerDied","Data":"a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6"} Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.449192 4818 scope.go:117] "RemoveContainer" containerID="8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.449117 4818 generic.go:334] "Generic (PLEG): container finished" podID="501083a5-d8f1-4171-8eca-c331c883ddb9" containerID="a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6" exitCode=0 Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.449315 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"501083a5-d8f1-4171-8eca-c331c883ddb9","Type":"ContainerDied","Data":"35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c"} Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.449243 4818 generic.go:334] "Generic (PLEG): container finished" podID="501083a5-d8f1-4171-8eca-c331c883ddb9" containerID="35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c" exitCode=0 Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.449415 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"501083a5-d8f1-4171-8eca-c331c883ddb9","Type":"ContainerDied","Data":"64816fd7802afcfb11caa78f0af1da023c2d1214034f77608782262c03620d55"} Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.449724 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0" containerName="nova-api-api" containerID="cri-o://fc69b2d8313b58d3dee5a61cbf5df5773080dd416463c96453502e8479bee49c" gracePeriod=30 Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.449744 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0" containerName="nova-api-log" containerID="cri-o://d20be45b2a4e0531fc49afd0ee6f16b2e30c5f3ab71110e1a37d338aedbdac59" gracePeriod=30 Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.501734 4818 scope.go:117] "RemoveContainer" containerID="3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.520195 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-ceilometer-tls-certs\") pod \"501083a5-d8f1-4171-8eca-c331c883ddb9\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.520233 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/501083a5-d8f1-4171-8eca-c331c883ddb9-log-httpd\") pod \"501083a5-d8f1-4171-8eca-c331c883ddb9\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.520329 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-combined-ca-bundle\") pod \"501083a5-d8f1-4171-8eca-c331c883ddb9\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.520378 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-config-data\") pod \"501083a5-d8f1-4171-8eca-c331c883ddb9\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.520482 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-scripts\") pod \"501083a5-d8f1-4171-8eca-c331c883ddb9\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.520594 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-sg-core-conf-yaml\") pod \"501083a5-d8f1-4171-8eca-c331c883ddb9\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.520663 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbdzq\" (UniqueName: \"kubernetes.io/projected/501083a5-d8f1-4171-8eca-c331c883ddb9-kube-api-access-gbdzq\") pod \"501083a5-d8f1-4171-8eca-c331c883ddb9\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.520711 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/501083a5-d8f1-4171-8eca-c331c883ddb9-run-httpd\") pod \"501083a5-d8f1-4171-8eca-c331c883ddb9\" (UID: \"501083a5-d8f1-4171-8eca-c331c883ddb9\") " Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.521315 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/501083a5-d8f1-4171-8eca-c331c883ddb9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "501083a5-d8f1-4171-8eca-c331c883ddb9" (UID: "501083a5-d8f1-4171-8eca-c331c883ddb9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.526586 4818 scope.go:117] "RemoveContainer" containerID="a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.527342 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/501083a5-d8f1-4171-8eca-c331c883ddb9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "501083a5-d8f1-4171-8eca-c331c883ddb9" (UID: "501083a5-d8f1-4171-8eca-c331c883ddb9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.530161 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-scripts" (OuterVolumeSpecName: "scripts") pod "501083a5-d8f1-4171-8eca-c331c883ddb9" (UID: "501083a5-d8f1-4171-8eca-c331c883ddb9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.532947 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/501083a5-d8f1-4171-8eca-c331c883ddb9-kube-api-access-gbdzq" (OuterVolumeSpecName: "kube-api-access-gbdzq") pod "501083a5-d8f1-4171-8eca-c331c883ddb9" (UID: "501083a5-d8f1-4171-8eca-c331c883ddb9"). InnerVolumeSpecName "kube-api-access-gbdzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.558846 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "501083a5-d8f1-4171-8eca-c331c883ddb9" (UID: "501083a5-d8f1-4171-8eca-c331c883ddb9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.595497 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "501083a5-d8f1-4171-8eca-c331c883ddb9" (UID: "501083a5-d8f1-4171-8eca-c331c883ddb9"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.625412 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.625471 4818 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.625483 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbdzq\" (UniqueName: \"kubernetes.io/projected/501083a5-d8f1-4171-8eca-c331c883ddb9-kube-api-access-gbdzq\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.625492 4818 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/501083a5-d8f1-4171-8eca-c331c883ddb9-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.625501 4818 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/501083a5-d8f1-4171-8eca-c331c883ddb9-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.625509 4818 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.639696 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "501083a5-d8f1-4171-8eca-c331c883ddb9" (UID: "501083a5-d8f1-4171-8eca-c331c883ddb9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.704423 4818 scope.go:117] "RemoveContainer" containerID="35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.727236 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.729387 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-config-data" (OuterVolumeSpecName: "config-data") pod "501083a5-d8f1-4171-8eca-c331c883ddb9" (UID: "501083a5-d8f1-4171-8eca-c331c883ddb9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.740579 4818 scope.go:117] "RemoveContainer" containerID="8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344" Nov 22 05:13:38 crc kubenswrapper[4818]: E1122 05:13:38.741105 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344\": container with ID starting with 8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344 not found: ID does not exist" containerID="8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.741145 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344"} err="failed to get container status \"8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344\": rpc error: code = NotFound desc = could not find container \"8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344\": container with ID starting with 8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344 not found: ID does not exist" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.741171 4818 scope.go:117] "RemoveContainer" containerID="3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2" Nov 22 05:13:38 crc kubenswrapper[4818]: E1122 05:13:38.741453 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2\": container with ID starting with 3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2 not found: ID does not exist" containerID="3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.741497 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2"} err="failed to get container status \"3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2\": rpc error: code = NotFound desc = could not find container \"3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2\": container with ID starting with 3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2 not found: ID does not exist" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.741527 4818 scope.go:117] "RemoveContainer" containerID="a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6" Nov 22 05:13:38 crc kubenswrapper[4818]: E1122 05:13:38.741824 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6\": container with ID starting with a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6 not found: ID does not exist" containerID="a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.741855 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6"} err="failed to get container status \"a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6\": rpc error: code = NotFound desc = could not find container \"a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6\": container with ID starting with a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6 not found: ID does not exist" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.741879 4818 scope.go:117] "RemoveContainer" containerID="35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c" Nov 22 05:13:38 crc kubenswrapper[4818]: E1122 05:13:38.742072 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c\": container with ID starting with 35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c not found: ID does not exist" containerID="35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.742099 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c"} err="failed to get container status \"35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c\": rpc error: code = NotFound desc = could not find container \"35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c\": container with ID starting with 35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c not found: ID does not exist" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.742115 4818 scope.go:117] "RemoveContainer" containerID="8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.742288 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344"} err="failed to get container status \"8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344\": rpc error: code = NotFound desc = could not find container \"8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344\": container with ID starting with 8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344 not found: ID does not exist" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.742310 4818 scope.go:117] "RemoveContainer" containerID="3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.742503 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2"} err="failed to get container status \"3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2\": rpc error: code = NotFound desc = could not find container \"3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2\": container with ID starting with 3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2 not found: ID does not exist" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.742522 4818 scope.go:117] "RemoveContainer" containerID="a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.742711 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6"} err="failed to get container status \"a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6\": rpc error: code = NotFound desc = could not find container \"a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6\": container with ID starting with a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6 not found: ID does not exist" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.742742 4818 scope.go:117] "RemoveContainer" containerID="35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.742941 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c"} err="failed to get container status \"35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c\": rpc error: code = NotFound desc = could not find container \"35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c\": container with ID starting with 35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c not found: ID does not exist" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.742960 4818 scope.go:117] "RemoveContainer" containerID="8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.743146 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344"} err="failed to get container status \"8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344\": rpc error: code = NotFound desc = could not find container \"8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344\": container with ID starting with 8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344 not found: ID does not exist" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.743174 4818 scope.go:117] "RemoveContainer" containerID="3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.743389 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2"} err="failed to get container status \"3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2\": rpc error: code = NotFound desc = could not find container \"3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2\": container with ID starting with 3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2 not found: ID does not exist" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.743418 4818 scope.go:117] "RemoveContainer" containerID="a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.743936 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6"} err="failed to get container status \"a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6\": rpc error: code = NotFound desc = could not find container \"a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6\": container with ID starting with a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6 not found: ID does not exist" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.743962 4818 scope.go:117] "RemoveContainer" containerID="35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.744478 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c"} err="failed to get container status \"35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c\": rpc error: code = NotFound desc = could not find container \"35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c\": container with ID starting with 35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c not found: ID does not exist" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.744502 4818 scope.go:117] "RemoveContainer" containerID="8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.744716 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344"} err="failed to get container status \"8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344\": rpc error: code = NotFound desc = could not find container \"8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344\": container with ID starting with 8c38703fcc5059dd511fee58046c6b1fe76efdfa0ae5638b99b4b7a64361f344 not found: ID does not exist" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.744741 4818 scope.go:117] "RemoveContainer" containerID="3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.744931 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2"} err="failed to get container status \"3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2\": rpc error: code = NotFound desc = could not find container \"3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2\": container with ID starting with 3d84f9c051dcf2a0841a1438cb0a9251f617b77c3bd4b5b10b2f9c53c06383a2 not found: ID does not exist" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.744951 4818 scope.go:117] "RemoveContainer" containerID="a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.745137 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6"} err="failed to get container status \"a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6\": rpc error: code = NotFound desc = could not find container \"a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6\": container with ID starting with a3363bd1194ad81b36822c5ba9509dcd5181df8cd84bed25ededaae5d4db82c6 not found: ID does not exist" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.745156 4818 scope.go:117] "RemoveContainer" containerID="35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.745373 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c"} err="failed to get container status \"35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c\": rpc error: code = NotFound desc = could not find container \"35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c\": container with ID starting with 35f114588c662b1f8442299980d4064996dd31246cf58ab18df0231be2d7937c not found: ID does not exist" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.807321 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b856c5697-88nx2" podStartSLOduration=3.807302807 podStartE2EDuration="3.807302807s" podCreationTimestamp="2025-11-22 05:13:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:13:38.462015495 +0000 UTC m=+1571.036432032" watchObservedRunningTime="2025-11-22 05:13:38.807302807 +0000 UTC m=+1571.381719334" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.808145 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.812389 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.822059 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.822138 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.822839 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.830848 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/501083a5-d8f1-4171-8eca-c331c883ddb9-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.830977 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:13:38 crc kubenswrapper[4818]: E1122 05:13:38.831484 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501083a5-d8f1-4171-8eca-c331c883ddb9" containerName="sg-core" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.831504 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="501083a5-d8f1-4171-8eca-c331c883ddb9" containerName="sg-core" Nov 22 05:13:38 crc kubenswrapper[4818]: E1122 05:13:38.831527 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501083a5-d8f1-4171-8eca-c331c883ddb9" containerName="ceilometer-notification-agent" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.831534 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="501083a5-d8f1-4171-8eca-c331c883ddb9" containerName="ceilometer-notification-agent" Nov 22 05:13:38 crc kubenswrapper[4818]: E1122 05:13:38.831547 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501083a5-d8f1-4171-8eca-c331c883ddb9" containerName="proxy-httpd" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.831554 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="501083a5-d8f1-4171-8eca-c331c883ddb9" containerName="proxy-httpd" Nov 22 05:13:38 crc kubenswrapper[4818]: E1122 05:13:38.831563 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501083a5-d8f1-4171-8eca-c331c883ddb9" containerName="ceilometer-central-agent" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.831569 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="501083a5-d8f1-4171-8eca-c331c883ddb9" containerName="ceilometer-central-agent" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.831750 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="501083a5-d8f1-4171-8eca-c331c883ddb9" containerName="ceilometer-central-agent" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.831767 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="501083a5-d8f1-4171-8eca-c331c883ddb9" containerName="sg-core" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.831776 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="501083a5-d8f1-4171-8eca-c331c883ddb9" containerName="proxy-httpd" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.831790 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="501083a5-d8f1-4171-8eca-c331c883ddb9" containerName="ceilometer-notification-agent" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.835161 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.837430 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.838387 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.841151 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.848430 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.932690 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/721b72d0-4e02-48cf-9fef-e49005870b5e-run-httpd\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.932819 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.933070 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-scripts\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.933094 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llfhb\" (UniqueName: \"kubernetes.io/projected/721b72d0-4e02-48cf-9fef-e49005870b5e-kube-api-access-llfhb\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.933136 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/721b72d0-4e02-48cf-9fef-e49005870b5e-log-httpd\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.933161 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-config-data\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.933181 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:38 crc kubenswrapper[4818]: I1122 05:13:38.933903 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:39 crc kubenswrapper[4818]: I1122 05:13:39.035143 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-scripts\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:39 crc kubenswrapper[4818]: I1122 05:13:39.035191 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llfhb\" (UniqueName: \"kubernetes.io/projected/721b72d0-4e02-48cf-9fef-e49005870b5e-kube-api-access-llfhb\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:39 crc kubenswrapper[4818]: I1122 05:13:39.035242 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/721b72d0-4e02-48cf-9fef-e49005870b5e-log-httpd\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:39 crc kubenswrapper[4818]: I1122 05:13:39.035298 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-config-data\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:39 crc kubenswrapper[4818]: I1122 05:13:39.035325 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:39 crc kubenswrapper[4818]: I1122 05:13:39.035369 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:39 crc kubenswrapper[4818]: I1122 05:13:39.035390 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/721b72d0-4e02-48cf-9fef-e49005870b5e-run-httpd\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:39 crc kubenswrapper[4818]: I1122 05:13:39.035450 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:39 crc kubenswrapper[4818]: I1122 05:13:39.036374 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/721b72d0-4e02-48cf-9fef-e49005870b5e-log-httpd\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:39 crc kubenswrapper[4818]: I1122 05:13:39.036691 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/721b72d0-4e02-48cf-9fef-e49005870b5e-run-httpd\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:39 crc kubenswrapper[4818]: I1122 05:13:39.038764 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:39 crc kubenswrapper[4818]: I1122 05:13:39.039052 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-scripts\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:39 crc kubenswrapper[4818]: I1122 05:13:39.042702 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:39 crc kubenswrapper[4818]: I1122 05:13:39.044287 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:39 crc kubenswrapper[4818]: I1122 05:13:39.044492 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-config-data\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:39 crc kubenswrapper[4818]: I1122 05:13:39.054198 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llfhb\" (UniqueName: \"kubernetes.io/projected/721b72d0-4e02-48cf-9fef-e49005870b5e-kube-api-access-llfhb\") pod \"ceilometer-0\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " pod="openstack/ceilometer-0" Nov 22 05:13:39 crc kubenswrapper[4818]: I1122 05:13:39.169238 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:13:39 crc kubenswrapper[4818]: I1122 05:13:39.463657 4818 generic.go:334] "Generic (PLEG): container finished" podID="5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0" containerID="d20be45b2a4e0531fc49afd0ee6f16b2e30c5f3ab71110e1a37d338aedbdac59" exitCode=143 Nov 22 05:13:39 crc kubenswrapper[4818]: I1122 05:13:39.464946 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0","Type":"ContainerDied","Data":"d20be45b2a4e0531fc49afd0ee6f16b2e30c5f3ab71110e1a37d338aedbdac59"} Nov 22 05:13:39 crc kubenswrapper[4818]: I1122 05:13:39.608712 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:13:40 crc kubenswrapper[4818]: I1122 05:13:40.066079 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:13:40 crc kubenswrapper[4818]: I1122 05:13:40.304187 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="501083a5-d8f1-4171-8eca-c331c883ddb9" path="/var/lib/kubelet/pods/501083a5-d8f1-4171-8eca-c331c883ddb9/volumes" Nov 22 05:13:40 crc kubenswrapper[4818]: I1122 05:13:40.477590 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"721b72d0-4e02-48cf-9fef-e49005870b5e","Type":"ContainerStarted","Data":"8050696ef62a14cf8a927b48692039ffd170abb14ab474134cb68e23b218e5e8"} Nov 22 05:13:40 crc kubenswrapper[4818]: I1122 05:13:40.477657 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"721b72d0-4e02-48cf-9fef-e49005870b5e","Type":"ContainerStarted","Data":"7a8dae4b162773f1233003426031781fda6e7df34243c7e9930c3ff456dcfb60"} Nov 22 05:13:41 crc kubenswrapper[4818]: I1122 05:13:41.501310 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"721b72d0-4e02-48cf-9fef-e49005870b5e","Type":"ContainerStarted","Data":"3dcfea0b5a849e2513b92df10fab605b3eb89f75d9094f1316a76be9da1e1d86"} Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.039569 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.194527 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-logs\") pod \"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0\" (UID: \"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0\") " Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.194584 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-config-data\") pod \"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0\" (UID: \"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0\") " Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.194682 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmnwr\" (UniqueName: \"kubernetes.io/projected/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-kube-api-access-tmnwr\") pod \"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0\" (UID: \"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0\") " Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.194826 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-combined-ca-bundle\") pod \"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0\" (UID: \"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0\") " Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.195346 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-logs" (OuterVolumeSpecName: "logs") pod "5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0" (UID: "5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.207593 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-kube-api-access-tmnwr" (OuterVolumeSpecName: "kube-api-access-tmnwr") pod "5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0" (UID: "5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0"). InnerVolumeSpecName "kube-api-access-tmnwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.223511 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-config-data" (OuterVolumeSpecName: "config-data") pod "5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0" (UID: "5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.239691 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0" (UID: "5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.298405 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-logs\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.298439 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.298454 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmnwr\" (UniqueName: \"kubernetes.io/projected/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-kube-api-access-tmnwr\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.298466 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.519881 4818 generic.go:334] "Generic (PLEG): container finished" podID="5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0" containerID="fc69b2d8313b58d3dee5a61cbf5df5773080dd416463c96453502e8479bee49c" exitCode=0 Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.519954 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.520018 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0","Type":"ContainerDied","Data":"fc69b2d8313b58d3dee5a61cbf5df5773080dd416463c96453502e8479bee49c"} Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.520054 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0","Type":"ContainerDied","Data":"46eb631546a92c46c242a7ed047cb5097613760c2ade558bf0acc9a57f584cfc"} Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.520077 4818 scope.go:117] "RemoveContainer" containerID="fc69b2d8313b58d3dee5a61cbf5df5773080dd416463c96453502e8479bee49c" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.528734 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"721b72d0-4e02-48cf-9fef-e49005870b5e","Type":"ContainerStarted","Data":"acd8bed780596986b7dc09ca74e96bdf9b78b821adee2a2dc57dccab14eead23"} Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.548424 4818 scope.go:117] "RemoveContainer" containerID="d20be45b2a4e0531fc49afd0ee6f16b2e30c5f3ab71110e1a37d338aedbdac59" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.551850 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.559846 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.584768 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 22 05:13:42 crc kubenswrapper[4818]: E1122 05:13:42.585178 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0" containerName="nova-api-api" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.585194 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0" containerName="nova-api-api" Nov 22 05:13:42 crc kubenswrapper[4818]: E1122 05:13:42.585228 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0" containerName="nova-api-log" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.585235 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0" containerName="nova-api-log" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.585405 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0" containerName="nova-api-api" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.585424 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0" containerName="nova-api-log" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.586330 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.587395 4818 scope.go:117] "RemoveContainer" containerID="fc69b2d8313b58d3dee5a61cbf5df5773080dd416463c96453502e8479bee49c" Nov 22 05:13:42 crc kubenswrapper[4818]: E1122 05:13:42.587914 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc69b2d8313b58d3dee5a61cbf5df5773080dd416463c96453502e8479bee49c\": container with ID starting with fc69b2d8313b58d3dee5a61cbf5df5773080dd416463c96453502e8479bee49c not found: ID does not exist" containerID="fc69b2d8313b58d3dee5a61cbf5df5773080dd416463c96453502e8479bee49c" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.587944 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc69b2d8313b58d3dee5a61cbf5df5773080dd416463c96453502e8479bee49c"} err="failed to get container status \"fc69b2d8313b58d3dee5a61cbf5df5773080dd416463c96453502e8479bee49c\": rpc error: code = NotFound desc = could not find container \"fc69b2d8313b58d3dee5a61cbf5df5773080dd416463c96453502e8479bee49c\": container with ID starting with fc69b2d8313b58d3dee5a61cbf5df5773080dd416463c96453502e8479bee49c not found: ID does not exist" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.587966 4818 scope.go:117] "RemoveContainer" containerID="d20be45b2a4e0531fc49afd0ee6f16b2e30c5f3ab71110e1a37d338aedbdac59" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.588644 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.588794 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.588864 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 22 05:13:42 crc kubenswrapper[4818]: E1122 05:13:42.592574 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d20be45b2a4e0531fc49afd0ee6f16b2e30c5f3ab71110e1a37d338aedbdac59\": container with ID starting with d20be45b2a4e0531fc49afd0ee6f16b2e30c5f3ab71110e1a37d338aedbdac59 not found: ID does not exist" containerID="d20be45b2a4e0531fc49afd0ee6f16b2e30c5f3ab71110e1a37d338aedbdac59" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.592646 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d20be45b2a4e0531fc49afd0ee6f16b2e30c5f3ab71110e1a37d338aedbdac59"} err="failed to get container status \"d20be45b2a4e0531fc49afd0ee6f16b2e30c5f3ab71110e1a37d338aedbdac59\": rpc error: code = NotFound desc = could not find container \"d20be45b2a4e0531fc49afd0ee6f16b2e30c5f3ab71110e1a37d338aedbdac59\": container with ID starting with d20be45b2a4e0531fc49afd0ee6f16b2e30c5f3ab71110e1a37d338aedbdac59 not found: ID does not exist" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.598215 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.712798 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-public-tls-certs\") pod \"nova-api-0\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.712872 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.712942 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-internal-tls-certs\") pod \"nova-api-0\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.713067 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb13e5c9-26fd-44d4-b745-b77ce47582ac-logs\") pod \"nova-api-0\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.713104 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-config-data\") pod \"nova-api-0\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.713156 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flnwx\" (UniqueName: \"kubernetes.io/projected/cb13e5c9-26fd-44d4-b745-b77ce47582ac-kube-api-access-flnwx\") pod \"nova-api-0\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.815094 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb13e5c9-26fd-44d4-b745-b77ce47582ac-logs\") pod \"nova-api-0\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.815178 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-config-data\") pod \"nova-api-0\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.815285 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flnwx\" (UniqueName: \"kubernetes.io/projected/cb13e5c9-26fd-44d4-b745-b77ce47582ac-kube-api-access-flnwx\") pod \"nova-api-0\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.815422 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-public-tls-certs\") pod \"nova-api-0\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.815464 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.815541 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-internal-tls-certs\") pod \"nova-api-0\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.816321 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb13e5c9-26fd-44d4-b745-b77ce47582ac-logs\") pod \"nova-api-0\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.820699 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.825055 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-internal-tls-certs\") pod \"nova-api-0\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.832167 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-public-tls-certs\") pod \"nova-api-0\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.834393 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-config-data\") pod \"nova-api-0\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.845012 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flnwx\" (UniqueName: \"kubernetes.io/projected/cb13e5c9-26fd-44d4-b745-b77ce47582ac-kube-api-access-flnwx\") pod \"nova-api-0\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " pod="openstack/nova-api-0" Nov 22 05:13:42 crc kubenswrapper[4818]: I1122 05:13:42.907076 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 05:13:43 crc kubenswrapper[4818]: W1122 05:13:43.386044 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb13e5c9_26fd_44d4_b745_b77ce47582ac.slice/crio-965c1341308b1275f5207a87cd376784fbc2da6264db1075c39dad63a60073f8 WatchSource:0}: Error finding container 965c1341308b1275f5207a87cd376784fbc2da6264db1075c39dad63a60073f8: Status 404 returned error can't find the container with id 965c1341308b1275f5207a87cd376784fbc2da6264db1075c39dad63a60073f8 Nov 22 05:13:43 crc kubenswrapper[4818]: I1122 05:13:43.387803 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 05:13:43 crc kubenswrapper[4818]: I1122 05:13:43.557768 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"721b72d0-4e02-48cf-9fef-e49005870b5e","Type":"ContainerStarted","Data":"5d60ec7181406bb7d24226231b7f8cc4b18581e96aa5016ae2f8b14c5ee53046"} Nov 22 05:13:43 crc kubenswrapper[4818]: I1122 05:13:43.557919 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="721b72d0-4e02-48cf-9fef-e49005870b5e" containerName="ceilometer-central-agent" containerID="cri-o://8050696ef62a14cf8a927b48692039ffd170abb14ab474134cb68e23b218e5e8" gracePeriod=30 Nov 22 05:13:43 crc kubenswrapper[4818]: I1122 05:13:43.558116 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 05:13:43 crc kubenswrapper[4818]: I1122 05:13:43.558514 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="721b72d0-4e02-48cf-9fef-e49005870b5e" containerName="proxy-httpd" containerID="cri-o://5d60ec7181406bb7d24226231b7f8cc4b18581e96aa5016ae2f8b14c5ee53046" gracePeriod=30 Nov 22 05:13:43 crc kubenswrapper[4818]: I1122 05:13:43.558578 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="721b72d0-4e02-48cf-9fef-e49005870b5e" containerName="sg-core" containerID="cri-o://acd8bed780596986b7dc09ca74e96bdf9b78b821adee2a2dc57dccab14eead23" gracePeriod=30 Nov 22 05:13:43 crc kubenswrapper[4818]: I1122 05:13:43.558620 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="721b72d0-4e02-48cf-9fef-e49005870b5e" containerName="ceilometer-notification-agent" containerID="cri-o://3dcfea0b5a849e2513b92df10fab605b3eb89f75d9094f1316a76be9da1e1d86" gracePeriod=30 Nov 22 05:13:43 crc kubenswrapper[4818]: I1122 05:13:43.565915 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cb13e5c9-26fd-44d4-b745-b77ce47582ac","Type":"ContainerStarted","Data":"965c1341308b1275f5207a87cd376784fbc2da6264db1075c39dad63a60073f8"} Nov 22 05:13:43 crc kubenswrapper[4818]: I1122 05:13:43.593996 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.048557654 podStartE2EDuration="5.593959716s" podCreationTimestamp="2025-11-22 05:13:38 +0000 UTC" firstStartedPulling="2025-11-22 05:13:39.619757396 +0000 UTC m=+1572.194173933" lastFinishedPulling="2025-11-22 05:13:43.165159468 +0000 UTC m=+1575.739575995" observedRunningTime="2025-11-22 05:13:43.587857524 +0000 UTC m=+1576.162274061" watchObservedRunningTime="2025-11-22 05:13:43.593959716 +0000 UTC m=+1576.168376233" Nov 22 05:13:43 crc kubenswrapper[4818]: E1122 05:13:43.723903 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod721b72d0_4e02_48cf_9fef_e49005870b5e.slice/crio-conmon-acd8bed780596986b7dc09ca74e96bdf9b78b821adee2a2dc57dccab14eead23.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod721b72d0_4e02_48cf_9fef_e49005870b5e.slice/crio-acd8bed780596986b7dc09ca74e96bdf9b78b821adee2a2dc57dccab14eead23.scope\": RecentStats: unable to find data in memory cache]" Nov 22 05:13:43 crc kubenswrapper[4818]: I1122 05:13:43.811604 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:43 crc kubenswrapper[4818]: I1122 05:13:43.820388 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 22 05:13:43 crc kubenswrapper[4818]: I1122 05:13:43.820507 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 22 05:13:43 crc kubenswrapper[4818]: I1122 05:13:43.830661 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.304976 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0" path="/var/lib/kubelet/pods/5998a0df-9ea3-4d81-92d8-01f4c9d7e2e0/volumes" Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.579548 4818 generic.go:334] "Generic (PLEG): container finished" podID="721b72d0-4e02-48cf-9fef-e49005870b5e" containerID="5d60ec7181406bb7d24226231b7f8cc4b18581e96aa5016ae2f8b14c5ee53046" exitCode=0 Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.579796 4818 generic.go:334] "Generic (PLEG): container finished" podID="721b72d0-4e02-48cf-9fef-e49005870b5e" containerID="acd8bed780596986b7dc09ca74e96bdf9b78b821adee2a2dc57dccab14eead23" exitCode=2 Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.579807 4818 generic.go:334] "Generic (PLEG): container finished" podID="721b72d0-4e02-48cf-9fef-e49005870b5e" containerID="3dcfea0b5a849e2513b92df10fab605b3eb89f75d9094f1316a76be9da1e1d86" exitCode=0 Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.579846 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"721b72d0-4e02-48cf-9fef-e49005870b5e","Type":"ContainerDied","Data":"5d60ec7181406bb7d24226231b7f8cc4b18581e96aa5016ae2f8b14c5ee53046"} Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.579873 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"721b72d0-4e02-48cf-9fef-e49005870b5e","Type":"ContainerDied","Data":"acd8bed780596986b7dc09ca74e96bdf9b78b821adee2a2dc57dccab14eead23"} Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.579886 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"721b72d0-4e02-48cf-9fef-e49005870b5e","Type":"ContainerDied","Data":"3dcfea0b5a849e2513b92df10fab605b3eb89f75d9094f1316a76be9da1e1d86"} Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.583161 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cb13e5c9-26fd-44d4-b745-b77ce47582ac","Type":"ContainerStarted","Data":"179c2ad91ce19d61606d928971b0be2664258d48a8a4b62a733db9e969aa162a"} Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.583188 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cb13e5c9-26fd-44d4-b745-b77ce47582ac","Type":"ContainerStarted","Data":"d43de0a0e44af675cab5f29fab0bfaae39db5c71d9d59bccd5b23fa8bb39aac3"} Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.616822 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.616801678 podStartE2EDuration="2.616801678s" podCreationTimestamp="2025-11-22 05:13:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:13:44.606870794 +0000 UTC m=+1577.181287331" watchObservedRunningTime="2025-11-22 05:13:44.616801678 +0000 UTC m=+1577.191218215" Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.631731 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.842695 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-lcvm7"] Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.846149 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lcvm7" Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.851096 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.851201 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.857312 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-lcvm7"] Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.857405 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.179:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.857423 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.179:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.973819 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdfdf908-61b8-4dcd-9898-fafdee790ca0-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lcvm7\" (UID: \"cdfdf908-61b8-4dcd-9898-fafdee790ca0\") " pod="openstack/nova-cell1-cell-mapping-lcvm7" Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.973876 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m46v4\" (UniqueName: \"kubernetes.io/projected/cdfdf908-61b8-4dcd-9898-fafdee790ca0-kube-api-access-m46v4\") pod \"nova-cell1-cell-mapping-lcvm7\" (UID: \"cdfdf908-61b8-4dcd-9898-fafdee790ca0\") " pod="openstack/nova-cell1-cell-mapping-lcvm7" Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.973932 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdfdf908-61b8-4dcd-9898-fafdee790ca0-scripts\") pod \"nova-cell1-cell-mapping-lcvm7\" (UID: \"cdfdf908-61b8-4dcd-9898-fafdee790ca0\") " pod="openstack/nova-cell1-cell-mapping-lcvm7" Nov 22 05:13:44 crc kubenswrapper[4818]: I1122 05:13:44.973954 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdfdf908-61b8-4dcd-9898-fafdee790ca0-config-data\") pod \"nova-cell1-cell-mapping-lcvm7\" (UID: \"cdfdf908-61b8-4dcd-9898-fafdee790ca0\") " pod="openstack/nova-cell1-cell-mapping-lcvm7" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.076852 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdfdf908-61b8-4dcd-9898-fafdee790ca0-config-data\") pod \"nova-cell1-cell-mapping-lcvm7\" (UID: \"cdfdf908-61b8-4dcd-9898-fafdee790ca0\") " pod="openstack/nova-cell1-cell-mapping-lcvm7" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.077396 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdfdf908-61b8-4dcd-9898-fafdee790ca0-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lcvm7\" (UID: \"cdfdf908-61b8-4dcd-9898-fafdee790ca0\") " pod="openstack/nova-cell1-cell-mapping-lcvm7" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.077609 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m46v4\" (UniqueName: \"kubernetes.io/projected/cdfdf908-61b8-4dcd-9898-fafdee790ca0-kube-api-access-m46v4\") pod \"nova-cell1-cell-mapping-lcvm7\" (UID: \"cdfdf908-61b8-4dcd-9898-fafdee790ca0\") " pod="openstack/nova-cell1-cell-mapping-lcvm7" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.077826 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdfdf908-61b8-4dcd-9898-fafdee790ca0-scripts\") pod \"nova-cell1-cell-mapping-lcvm7\" (UID: \"cdfdf908-61b8-4dcd-9898-fafdee790ca0\") " pod="openstack/nova-cell1-cell-mapping-lcvm7" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.084352 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdfdf908-61b8-4dcd-9898-fafdee790ca0-config-data\") pod \"nova-cell1-cell-mapping-lcvm7\" (UID: \"cdfdf908-61b8-4dcd-9898-fafdee790ca0\") " pod="openstack/nova-cell1-cell-mapping-lcvm7" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.088665 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdfdf908-61b8-4dcd-9898-fafdee790ca0-scripts\") pod \"nova-cell1-cell-mapping-lcvm7\" (UID: \"cdfdf908-61b8-4dcd-9898-fafdee790ca0\") " pod="openstack/nova-cell1-cell-mapping-lcvm7" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.089896 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdfdf908-61b8-4dcd-9898-fafdee790ca0-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lcvm7\" (UID: \"cdfdf908-61b8-4dcd-9898-fafdee790ca0\") " pod="openstack/nova-cell1-cell-mapping-lcvm7" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.101444 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m46v4\" (UniqueName: \"kubernetes.io/projected/cdfdf908-61b8-4dcd-9898-fafdee790ca0-kube-api-access-m46v4\") pod \"nova-cell1-cell-mapping-lcvm7\" (UID: \"cdfdf908-61b8-4dcd-9898-fafdee790ca0\") " pod="openstack/nova-cell1-cell-mapping-lcvm7" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.175050 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lcvm7" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.185271 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.280506 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-scripts\") pod \"721b72d0-4e02-48cf-9fef-e49005870b5e\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.280841 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-config-data\") pod \"721b72d0-4e02-48cf-9fef-e49005870b5e\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.280917 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llfhb\" (UniqueName: \"kubernetes.io/projected/721b72d0-4e02-48cf-9fef-e49005870b5e-kube-api-access-llfhb\") pod \"721b72d0-4e02-48cf-9fef-e49005870b5e\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.280990 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-ceilometer-tls-certs\") pod \"721b72d0-4e02-48cf-9fef-e49005870b5e\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.281042 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/721b72d0-4e02-48cf-9fef-e49005870b5e-run-httpd\") pod \"721b72d0-4e02-48cf-9fef-e49005870b5e\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.281066 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-combined-ca-bundle\") pod \"721b72d0-4e02-48cf-9fef-e49005870b5e\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.281125 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-sg-core-conf-yaml\") pod \"721b72d0-4e02-48cf-9fef-e49005870b5e\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.281186 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/721b72d0-4e02-48cf-9fef-e49005870b5e-log-httpd\") pod \"721b72d0-4e02-48cf-9fef-e49005870b5e\" (UID: \"721b72d0-4e02-48cf-9fef-e49005870b5e\") " Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.282238 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/721b72d0-4e02-48cf-9fef-e49005870b5e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "721b72d0-4e02-48cf-9fef-e49005870b5e" (UID: "721b72d0-4e02-48cf-9fef-e49005870b5e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.285886 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/721b72d0-4e02-48cf-9fef-e49005870b5e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "721b72d0-4e02-48cf-9fef-e49005870b5e" (UID: "721b72d0-4e02-48cf-9fef-e49005870b5e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.288206 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-scripts" (OuterVolumeSpecName: "scripts") pod "721b72d0-4e02-48cf-9fef-e49005870b5e" (UID: "721b72d0-4e02-48cf-9fef-e49005870b5e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.291804 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/721b72d0-4e02-48cf-9fef-e49005870b5e-kube-api-access-llfhb" (OuterVolumeSpecName: "kube-api-access-llfhb") pod "721b72d0-4e02-48cf-9fef-e49005870b5e" (UID: "721b72d0-4e02-48cf-9fef-e49005870b5e"). InnerVolumeSpecName "kube-api-access-llfhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.310912 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "721b72d0-4e02-48cf-9fef-e49005870b5e" (UID: "721b72d0-4e02-48cf-9fef-e49005870b5e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.363504 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "721b72d0-4e02-48cf-9fef-e49005870b5e" (UID: "721b72d0-4e02-48cf-9fef-e49005870b5e"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.383279 4818 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.383311 4818 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/721b72d0-4e02-48cf-9fef-e49005870b5e-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.383319 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.383330 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llfhb\" (UniqueName: \"kubernetes.io/projected/721b72d0-4e02-48cf-9fef-e49005870b5e-kube-api-access-llfhb\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.383340 4818 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.383348 4818 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/721b72d0-4e02-48cf-9fef-e49005870b5e-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.406549 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "721b72d0-4e02-48cf-9fef-e49005870b5e" (UID: "721b72d0-4e02-48cf-9fef-e49005870b5e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.411158 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-config-data" (OuterVolumeSpecName: "config-data") pod "721b72d0-4e02-48cf-9fef-e49005870b5e" (UID: "721b72d0-4e02-48cf-9fef-e49005870b5e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.485222 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.485270 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/721b72d0-4e02-48cf-9fef-e49005870b5e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.600313 4818 generic.go:334] "Generic (PLEG): container finished" podID="721b72d0-4e02-48cf-9fef-e49005870b5e" containerID="8050696ef62a14cf8a927b48692039ffd170abb14ab474134cb68e23b218e5e8" exitCode=0 Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.601797 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.610654 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"721b72d0-4e02-48cf-9fef-e49005870b5e","Type":"ContainerDied","Data":"8050696ef62a14cf8a927b48692039ffd170abb14ab474134cb68e23b218e5e8"} Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.610713 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"721b72d0-4e02-48cf-9fef-e49005870b5e","Type":"ContainerDied","Data":"7a8dae4b162773f1233003426031781fda6e7df34243c7e9930c3ff456dcfb60"} Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.610731 4818 scope.go:117] "RemoveContainer" containerID="5d60ec7181406bb7d24226231b7f8cc4b18581e96aa5016ae2f8b14c5ee53046" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.644026 4818 scope.go:117] "RemoveContainer" containerID="acd8bed780596986b7dc09ca74e96bdf9b78b821adee2a2dc57dccab14eead23" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.673983 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.681733 4818 scope.go:117] "RemoveContainer" containerID="3dcfea0b5a849e2513b92df10fab605b3eb89f75d9094f1316a76be9da1e1d86" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.684237 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-lcvm7"] Nov 22 05:13:45 crc kubenswrapper[4818]: W1122 05:13:45.690469 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdfdf908_61b8_4dcd_9898_fafdee790ca0.slice/crio-3a210a43f18e56bf8f05964f844cba4c5bd8eff0d5ac6a7cb1261c23a69d1fd7 WatchSource:0}: Error finding container 3a210a43f18e56bf8f05964f844cba4c5bd8eff0d5ac6a7cb1261c23a69d1fd7: Status 404 returned error can't find the container with id 3a210a43f18e56bf8f05964f844cba4c5bd8eff0d5ac6a7cb1261c23a69d1fd7 Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.702887 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.713890 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:13:45 crc kubenswrapper[4818]: E1122 05:13:45.714500 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="721b72d0-4e02-48cf-9fef-e49005870b5e" containerName="ceilometer-central-agent" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.714523 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="721b72d0-4e02-48cf-9fef-e49005870b5e" containerName="ceilometer-central-agent" Nov 22 05:13:45 crc kubenswrapper[4818]: E1122 05:13:45.714539 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="721b72d0-4e02-48cf-9fef-e49005870b5e" containerName="proxy-httpd" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.714549 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="721b72d0-4e02-48cf-9fef-e49005870b5e" containerName="proxy-httpd" Nov 22 05:13:45 crc kubenswrapper[4818]: E1122 05:13:45.714566 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="721b72d0-4e02-48cf-9fef-e49005870b5e" containerName="ceilometer-notification-agent" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.714576 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="721b72d0-4e02-48cf-9fef-e49005870b5e" containerName="ceilometer-notification-agent" Nov 22 05:13:45 crc kubenswrapper[4818]: E1122 05:13:45.714603 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="721b72d0-4e02-48cf-9fef-e49005870b5e" containerName="sg-core" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.714611 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="721b72d0-4e02-48cf-9fef-e49005870b5e" containerName="sg-core" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.714830 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="721b72d0-4e02-48cf-9fef-e49005870b5e" containerName="ceilometer-notification-agent" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.714846 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="721b72d0-4e02-48cf-9fef-e49005870b5e" containerName="sg-core" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.714881 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="721b72d0-4e02-48cf-9fef-e49005870b5e" containerName="proxy-httpd" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.714891 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="721b72d0-4e02-48cf-9fef-e49005870b5e" containerName="ceilometer-central-agent" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.719092 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.736512 4818 scope.go:117] "RemoveContainer" containerID="8050696ef62a14cf8a927b48692039ffd170abb14ab474134cb68e23b218e5e8" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.744060 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.745989 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.746576 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.747656 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.784056 4818 scope.go:117] "RemoveContainer" containerID="5d60ec7181406bb7d24226231b7f8cc4b18581e96aa5016ae2f8b14c5ee53046" Nov 22 05:13:45 crc kubenswrapper[4818]: E1122 05:13:45.784684 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d60ec7181406bb7d24226231b7f8cc4b18581e96aa5016ae2f8b14c5ee53046\": container with ID starting with 5d60ec7181406bb7d24226231b7f8cc4b18581e96aa5016ae2f8b14c5ee53046 not found: ID does not exist" containerID="5d60ec7181406bb7d24226231b7f8cc4b18581e96aa5016ae2f8b14c5ee53046" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.784731 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d60ec7181406bb7d24226231b7f8cc4b18581e96aa5016ae2f8b14c5ee53046"} err="failed to get container status \"5d60ec7181406bb7d24226231b7f8cc4b18581e96aa5016ae2f8b14c5ee53046\": rpc error: code = NotFound desc = could not find container \"5d60ec7181406bb7d24226231b7f8cc4b18581e96aa5016ae2f8b14c5ee53046\": container with ID starting with 5d60ec7181406bb7d24226231b7f8cc4b18581e96aa5016ae2f8b14c5ee53046 not found: ID does not exist" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.784751 4818 scope.go:117] "RemoveContainer" containerID="acd8bed780596986b7dc09ca74e96bdf9b78b821adee2a2dc57dccab14eead23" Nov 22 05:13:45 crc kubenswrapper[4818]: E1122 05:13:45.785054 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acd8bed780596986b7dc09ca74e96bdf9b78b821adee2a2dc57dccab14eead23\": container with ID starting with acd8bed780596986b7dc09ca74e96bdf9b78b821adee2a2dc57dccab14eead23 not found: ID does not exist" containerID="acd8bed780596986b7dc09ca74e96bdf9b78b821adee2a2dc57dccab14eead23" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.785081 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acd8bed780596986b7dc09ca74e96bdf9b78b821adee2a2dc57dccab14eead23"} err="failed to get container status \"acd8bed780596986b7dc09ca74e96bdf9b78b821adee2a2dc57dccab14eead23\": rpc error: code = NotFound desc = could not find container \"acd8bed780596986b7dc09ca74e96bdf9b78b821adee2a2dc57dccab14eead23\": container with ID starting with acd8bed780596986b7dc09ca74e96bdf9b78b821adee2a2dc57dccab14eead23 not found: ID does not exist" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.785098 4818 scope.go:117] "RemoveContainer" containerID="3dcfea0b5a849e2513b92df10fab605b3eb89f75d9094f1316a76be9da1e1d86" Nov 22 05:13:45 crc kubenswrapper[4818]: E1122 05:13:45.785430 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dcfea0b5a849e2513b92df10fab605b3eb89f75d9094f1316a76be9da1e1d86\": container with ID starting with 3dcfea0b5a849e2513b92df10fab605b3eb89f75d9094f1316a76be9da1e1d86 not found: ID does not exist" containerID="3dcfea0b5a849e2513b92df10fab605b3eb89f75d9094f1316a76be9da1e1d86" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.785486 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dcfea0b5a849e2513b92df10fab605b3eb89f75d9094f1316a76be9da1e1d86"} err="failed to get container status \"3dcfea0b5a849e2513b92df10fab605b3eb89f75d9094f1316a76be9da1e1d86\": rpc error: code = NotFound desc = could not find container \"3dcfea0b5a849e2513b92df10fab605b3eb89f75d9094f1316a76be9da1e1d86\": container with ID starting with 3dcfea0b5a849e2513b92df10fab605b3eb89f75d9094f1316a76be9da1e1d86 not found: ID does not exist" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.785556 4818 scope.go:117] "RemoveContainer" containerID="8050696ef62a14cf8a927b48692039ffd170abb14ab474134cb68e23b218e5e8" Nov 22 05:13:45 crc kubenswrapper[4818]: E1122 05:13:45.785817 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8050696ef62a14cf8a927b48692039ffd170abb14ab474134cb68e23b218e5e8\": container with ID starting with 8050696ef62a14cf8a927b48692039ffd170abb14ab474134cb68e23b218e5e8 not found: ID does not exist" containerID="8050696ef62a14cf8a927b48692039ffd170abb14ab474134cb68e23b218e5e8" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.785843 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8050696ef62a14cf8a927b48692039ffd170abb14ab474134cb68e23b218e5e8"} err="failed to get container status \"8050696ef62a14cf8a927b48692039ffd170abb14ab474134cb68e23b218e5e8\": rpc error: code = NotFound desc = could not find container \"8050696ef62a14cf8a927b48692039ffd170abb14ab474134cb68e23b218e5e8\": container with ID starting with 8050696ef62a14cf8a927b48692039ffd170abb14ab474134cb68e23b218e5e8 not found: ID does not exist" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.892243 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-log-httpd\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.892337 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.892372 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cdlm\" (UniqueName: \"kubernetes.io/projected/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-kube-api-access-7cdlm\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.892449 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.892479 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-run-httpd\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.892518 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.892551 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-scripts\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.892626 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-config-data\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.932428 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.994629 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-config-data\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.994729 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-log-httpd\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.994777 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.994801 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cdlm\" (UniqueName: \"kubernetes.io/projected/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-kube-api-access-7cdlm\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.994896 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.994942 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-run-httpd\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.994970 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.995021 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-scripts\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:45 crc kubenswrapper[4818]: I1122 05:13:45.995920 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-run-httpd\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:45.999871 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.000399 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-log-httpd\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.001290 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-wl6kj"] Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.001500 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" podUID="34665b87-c9bb-4b11-9406-57bc79ed2e0c" containerName="dnsmasq-dns" containerID="cri-o://3bbe4c52bc939b728f48822e01bec5b824c46735ebe31147ec275e3f4aa58476" gracePeriod=10 Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.001574 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-config-data\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.007957 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.008758 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-scripts\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.008935 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.020345 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cdlm\" (UniqueName: \"kubernetes.io/projected/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-kube-api-access-7cdlm\") pod \"ceilometer-0\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " pod="openstack/ceilometer-0" Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.072027 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.304365 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="721b72d0-4e02-48cf-9fef-e49005870b5e" path="/var/lib/kubelet/pods/721b72d0-4e02-48cf-9fef-e49005870b5e/volumes" Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.541985 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.630394 4818 generic.go:334] "Generic (PLEG): container finished" podID="34665b87-c9bb-4b11-9406-57bc79ed2e0c" containerID="3bbe4c52bc939b728f48822e01bec5b824c46735ebe31147ec275e3f4aa58476" exitCode=0 Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.630454 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" event={"ID":"34665b87-c9bb-4b11-9406-57bc79ed2e0c","Type":"ContainerDied","Data":"3bbe4c52bc939b728f48822e01bec5b824c46735ebe31147ec275e3f4aa58476"} Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.631715 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b","Type":"ContainerStarted","Data":"9e4c64805b1f7cbcf57a08669bc330fafdbd9272783cd0780077446e44e413f5"} Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.633446 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lcvm7" event={"ID":"cdfdf908-61b8-4dcd-9898-fafdee790ca0","Type":"ContainerStarted","Data":"42c56e4d11f063f9ca8ed3cfef725446df4c6f534b616e9dbc311be53c36d939"} Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.633491 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lcvm7" event={"ID":"cdfdf908-61b8-4dcd-9898-fafdee790ca0","Type":"ContainerStarted","Data":"3a210a43f18e56bf8f05964f844cba4c5bd8eff0d5ac6a7cb1261c23a69d1fd7"} Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.654705 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-lcvm7" podStartSLOduration=2.654684295 podStartE2EDuration="2.654684295s" podCreationTimestamp="2025-11-22 05:13:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:13:46.646314243 +0000 UTC m=+1579.220730770" watchObservedRunningTime="2025-11-22 05:13:46.654684295 +0000 UTC m=+1579.229100822" Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.818071 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.913075 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-config\") pod \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\" (UID: \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\") " Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.913123 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7f4lc\" (UniqueName: \"kubernetes.io/projected/34665b87-c9bb-4b11-9406-57bc79ed2e0c-kube-api-access-7f4lc\") pod \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\" (UID: \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\") " Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.913156 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-ovsdbserver-sb\") pod \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\" (UID: \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\") " Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.913343 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-ovsdbserver-nb\") pod \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\" (UID: \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\") " Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.913388 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-dns-svc\") pod \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\" (UID: \"34665b87-c9bb-4b11-9406-57bc79ed2e0c\") " Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.923721 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34665b87-c9bb-4b11-9406-57bc79ed2e0c-kube-api-access-7f4lc" (OuterVolumeSpecName: "kube-api-access-7f4lc") pod "34665b87-c9bb-4b11-9406-57bc79ed2e0c" (UID: "34665b87-c9bb-4b11-9406-57bc79ed2e0c"). InnerVolumeSpecName "kube-api-access-7f4lc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.957942 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "34665b87-c9bb-4b11-9406-57bc79ed2e0c" (UID: "34665b87-c9bb-4b11-9406-57bc79ed2e0c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.965149 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-config" (OuterVolumeSpecName: "config") pod "34665b87-c9bb-4b11-9406-57bc79ed2e0c" (UID: "34665b87-c9bb-4b11-9406-57bc79ed2e0c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.977076 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "34665b87-c9bb-4b11-9406-57bc79ed2e0c" (UID: "34665b87-c9bb-4b11-9406-57bc79ed2e0c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:13:46 crc kubenswrapper[4818]: I1122 05:13:46.993445 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "34665b87-c9bb-4b11-9406-57bc79ed2e0c" (UID: "34665b87-c9bb-4b11-9406-57bc79ed2e0c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:13:47 crc kubenswrapper[4818]: I1122 05:13:47.016405 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:47 crc kubenswrapper[4818]: I1122 05:13:47.016437 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7f4lc\" (UniqueName: \"kubernetes.io/projected/34665b87-c9bb-4b11-9406-57bc79ed2e0c-kube-api-access-7f4lc\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:47 crc kubenswrapper[4818]: I1122 05:13:47.016451 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:47 crc kubenswrapper[4818]: I1122 05:13:47.016459 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:47 crc kubenswrapper[4818]: I1122 05:13:47.016469 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34665b87-c9bb-4b11-9406-57bc79ed2e0c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:47 crc kubenswrapper[4818]: I1122 05:13:47.655895 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" event={"ID":"34665b87-c9bb-4b11-9406-57bc79ed2e0c","Type":"ContainerDied","Data":"4f6898013226a2a6e98093db4154ae814ce19b76449da3c797b00ce7b3bdd899"} Nov 22 05:13:47 crc kubenswrapper[4818]: I1122 05:13:47.656112 4818 scope.go:117] "RemoveContainer" containerID="3bbe4c52bc939b728f48822e01bec5b824c46735ebe31147ec275e3f4aa58476" Nov 22 05:13:47 crc kubenswrapper[4818]: I1122 05:13:47.656294 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-wl6kj" Nov 22 05:13:47 crc kubenswrapper[4818]: I1122 05:13:47.684375 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b","Type":"ContainerStarted","Data":"91ce5cbe3ab722e8bc8fdc17560fba78094af6ccf82882c2e3c70847e57ed074"} Nov 22 05:13:47 crc kubenswrapper[4818]: I1122 05:13:47.711164 4818 scope.go:117] "RemoveContainer" containerID="441378f5a899b66f84bdf03159b3460c7aebc5518fd93ee42ae52eea9f2ae661" Nov 22 05:13:47 crc kubenswrapper[4818]: I1122 05:13:47.716883 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-wl6kj"] Nov 22 05:13:47 crc kubenswrapper[4818]: I1122 05:13:47.739766 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-wl6kj"] Nov 22 05:13:48 crc kubenswrapper[4818]: I1122 05:13:48.305278 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34665b87-c9bb-4b11-9406-57bc79ed2e0c" path="/var/lib/kubelet/pods/34665b87-c9bb-4b11-9406-57bc79ed2e0c/volumes" Nov 22 05:13:48 crc kubenswrapper[4818]: I1122 05:13:48.698320 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b","Type":"ContainerStarted","Data":"319960ceed8832f65e4612ecc57252b581b471b3838923554855f0a059d72e2c"} Nov 22 05:13:48 crc kubenswrapper[4818]: I1122 05:13:48.698790 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b","Type":"ContainerStarted","Data":"d8a0f61bcb26e6c51fa6223cd0bff31d90f370eddb5ed5988e7c1af901ec93b9"} Nov 22 05:13:50 crc kubenswrapper[4818]: I1122 05:13:50.722778 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b","Type":"ContainerStarted","Data":"634cc8f12feb62b2d0d5acc75244240144520c1342018fc39725c9df8ce14ecb"} Nov 22 05:13:50 crc kubenswrapper[4818]: I1122 05:13:50.723163 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 05:13:50 crc kubenswrapper[4818]: I1122 05:13:50.725129 4818 generic.go:334] "Generic (PLEG): container finished" podID="cdfdf908-61b8-4dcd-9898-fafdee790ca0" containerID="42c56e4d11f063f9ca8ed3cfef725446df4c6f534b616e9dbc311be53c36d939" exitCode=0 Nov 22 05:13:50 crc kubenswrapper[4818]: I1122 05:13:50.725182 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lcvm7" event={"ID":"cdfdf908-61b8-4dcd-9898-fafdee790ca0","Type":"ContainerDied","Data":"42c56e4d11f063f9ca8ed3cfef725446df4c6f534b616e9dbc311be53c36d939"} Nov 22 05:13:50 crc kubenswrapper[4818]: I1122 05:13:50.761398 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.694001055 podStartE2EDuration="5.761380901s" podCreationTimestamp="2025-11-22 05:13:45 +0000 UTC" firstStartedPulling="2025-11-22 05:13:46.569642009 +0000 UTC m=+1579.144058536" lastFinishedPulling="2025-11-22 05:13:49.637021845 +0000 UTC m=+1582.211438382" observedRunningTime="2025-11-22 05:13:50.756997165 +0000 UTC m=+1583.331413702" watchObservedRunningTime="2025-11-22 05:13:50.761380901 +0000 UTC m=+1583.335797428" Nov 22 05:13:51 crc kubenswrapper[4818]: I1122 05:13:51.265966 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:13:51 crc kubenswrapper[4818]: I1122 05:13:51.266143 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:13:52 crc kubenswrapper[4818]: I1122 05:13:52.096702 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lcvm7" Nov 22 05:13:52 crc kubenswrapper[4818]: I1122 05:13:52.227501 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdfdf908-61b8-4dcd-9898-fafdee790ca0-config-data\") pod \"cdfdf908-61b8-4dcd-9898-fafdee790ca0\" (UID: \"cdfdf908-61b8-4dcd-9898-fafdee790ca0\") " Nov 22 05:13:52 crc kubenswrapper[4818]: I1122 05:13:52.227680 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdfdf908-61b8-4dcd-9898-fafdee790ca0-scripts\") pod \"cdfdf908-61b8-4dcd-9898-fafdee790ca0\" (UID: \"cdfdf908-61b8-4dcd-9898-fafdee790ca0\") " Nov 22 05:13:52 crc kubenswrapper[4818]: I1122 05:13:52.227816 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m46v4\" (UniqueName: \"kubernetes.io/projected/cdfdf908-61b8-4dcd-9898-fafdee790ca0-kube-api-access-m46v4\") pod \"cdfdf908-61b8-4dcd-9898-fafdee790ca0\" (UID: \"cdfdf908-61b8-4dcd-9898-fafdee790ca0\") " Nov 22 05:13:52 crc kubenswrapper[4818]: I1122 05:13:52.228075 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdfdf908-61b8-4dcd-9898-fafdee790ca0-combined-ca-bundle\") pod \"cdfdf908-61b8-4dcd-9898-fafdee790ca0\" (UID: \"cdfdf908-61b8-4dcd-9898-fafdee790ca0\") " Nov 22 05:13:52 crc kubenswrapper[4818]: I1122 05:13:52.236821 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdfdf908-61b8-4dcd-9898-fafdee790ca0-kube-api-access-m46v4" (OuterVolumeSpecName: "kube-api-access-m46v4") pod "cdfdf908-61b8-4dcd-9898-fafdee790ca0" (UID: "cdfdf908-61b8-4dcd-9898-fafdee790ca0"). InnerVolumeSpecName "kube-api-access-m46v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:13:52 crc kubenswrapper[4818]: I1122 05:13:52.237903 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdfdf908-61b8-4dcd-9898-fafdee790ca0-scripts" (OuterVolumeSpecName: "scripts") pod "cdfdf908-61b8-4dcd-9898-fafdee790ca0" (UID: "cdfdf908-61b8-4dcd-9898-fafdee790ca0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:52 crc kubenswrapper[4818]: I1122 05:13:52.263357 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdfdf908-61b8-4dcd-9898-fafdee790ca0-config-data" (OuterVolumeSpecName: "config-data") pod "cdfdf908-61b8-4dcd-9898-fafdee790ca0" (UID: "cdfdf908-61b8-4dcd-9898-fafdee790ca0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:52 crc kubenswrapper[4818]: I1122 05:13:52.263920 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdfdf908-61b8-4dcd-9898-fafdee790ca0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cdfdf908-61b8-4dcd-9898-fafdee790ca0" (UID: "cdfdf908-61b8-4dcd-9898-fafdee790ca0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:52 crc kubenswrapper[4818]: I1122 05:13:52.330413 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdfdf908-61b8-4dcd-9898-fafdee790ca0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:52 crc kubenswrapper[4818]: I1122 05:13:52.330450 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdfdf908-61b8-4dcd-9898-fafdee790ca0-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:52 crc kubenswrapper[4818]: I1122 05:13:52.330458 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdfdf908-61b8-4dcd-9898-fafdee790ca0-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:52 crc kubenswrapper[4818]: I1122 05:13:52.330467 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m46v4\" (UniqueName: \"kubernetes.io/projected/cdfdf908-61b8-4dcd-9898-fafdee790ca0-kube-api-access-m46v4\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:52 crc kubenswrapper[4818]: I1122 05:13:52.744901 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lcvm7" event={"ID":"cdfdf908-61b8-4dcd-9898-fafdee790ca0","Type":"ContainerDied","Data":"3a210a43f18e56bf8f05964f844cba4c5bd8eff0d5ac6a7cb1261c23a69d1fd7"} Nov 22 05:13:52 crc kubenswrapper[4818]: I1122 05:13:52.744944 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a210a43f18e56bf8f05964f844cba4c5bd8eff0d5ac6a7cb1261c23a69d1fd7" Nov 22 05:13:52 crc kubenswrapper[4818]: I1122 05:13:52.744999 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lcvm7" Nov 22 05:13:52 crc kubenswrapper[4818]: I1122 05:13:52.907898 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 05:13:52 crc kubenswrapper[4818]: I1122 05:13:52.907972 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 05:13:53 crc kubenswrapper[4818]: I1122 05:13:53.062939 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 05:13:53 crc kubenswrapper[4818]: I1122 05:13:53.063332 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="493288a9-105f-4ac3-8f6e-2fde520d3d19" containerName="nova-scheduler-scheduler" containerID="cri-o://4f6b54fda6fd2e611e1267b94b1d33df9e78737c66fdc769a1fb78e46a2fc845" gracePeriod=30 Nov 22 05:13:53 crc kubenswrapper[4818]: I1122 05:13:53.077834 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 22 05:13:53 crc kubenswrapper[4818]: I1122 05:13:53.126954 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 05:13:53 crc kubenswrapper[4818]: I1122 05:13:53.127187 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12" containerName="nova-metadata-log" containerID="cri-o://291f8def5fd201497ecca5e29c50e6eed2daff4a131454883b80d38d73a92927" gracePeriod=30 Nov 22 05:13:53 crc kubenswrapper[4818]: I1122 05:13:53.127331 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12" containerName="nova-metadata-metadata" containerID="cri-o://51dd0c67d382bf1b8bbe35e6cac4fe11396139cdb92c92f82e2aeab8dcf57231" gracePeriod=30 Nov 22 05:13:53 crc kubenswrapper[4818]: I1122 05:13:53.754561 4818 generic.go:334] "Generic (PLEG): container finished" podID="7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12" containerID="291f8def5fd201497ecca5e29c50e6eed2daff4a131454883b80d38d73a92927" exitCode=143 Nov 22 05:13:53 crc kubenswrapper[4818]: I1122 05:13:53.754618 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12","Type":"ContainerDied","Data":"291f8def5fd201497ecca5e29c50e6eed2daff4a131454883b80d38d73a92927"} Nov 22 05:13:53 crc kubenswrapper[4818]: I1122 05:13:53.754804 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cb13e5c9-26fd-44d4-b745-b77ce47582ac" containerName="nova-api-log" containerID="cri-o://d43de0a0e44af675cab5f29fab0bfaae39db5c71d9d59bccd5b23fa8bb39aac3" gracePeriod=30 Nov 22 05:13:53 crc kubenswrapper[4818]: I1122 05:13:53.754886 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cb13e5c9-26fd-44d4-b745-b77ce47582ac" containerName="nova-api-api" containerID="cri-o://179c2ad91ce19d61606d928971b0be2664258d48a8a4b62a733db9e969aa162a" gracePeriod=30 Nov 22 05:13:53 crc kubenswrapper[4818]: I1122 05:13:53.768865 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cb13e5c9-26fd-44d4-b745-b77ce47582ac" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.182:8774/\": EOF" Nov 22 05:13:53 crc kubenswrapper[4818]: I1122 05:13:53.769476 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cb13e5c9-26fd-44d4-b745-b77ce47582ac" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.182:8774/\": EOF" Nov 22 05:13:53 crc kubenswrapper[4818]: E1122 05:13:53.975238 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb13e5c9_26fd_44d4_b745_b77ce47582ac.slice/crio-conmon-d43de0a0e44af675cab5f29fab0bfaae39db5c71d9d59bccd5b23fa8bb39aac3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb13e5c9_26fd_44d4_b745_b77ce47582ac.slice/crio-d43de0a0e44af675cab5f29fab0bfaae39db5c71d9d59bccd5b23fa8bb39aac3.scope\": RecentStats: unable to find data in memory cache]" Nov 22 05:13:54 crc kubenswrapper[4818]: I1122 05:13:54.758461 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 05:13:54 crc kubenswrapper[4818]: I1122 05:13:54.770881 4818 generic.go:334] "Generic (PLEG): container finished" podID="cb13e5c9-26fd-44d4-b745-b77ce47582ac" containerID="d43de0a0e44af675cab5f29fab0bfaae39db5c71d9d59bccd5b23fa8bb39aac3" exitCode=143 Nov 22 05:13:54 crc kubenswrapper[4818]: I1122 05:13:54.770915 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cb13e5c9-26fd-44d4-b745-b77ce47582ac","Type":"ContainerDied","Data":"d43de0a0e44af675cab5f29fab0bfaae39db5c71d9d59bccd5b23fa8bb39aac3"} Nov 22 05:13:54 crc kubenswrapper[4818]: I1122 05:13:54.772913 4818 generic.go:334] "Generic (PLEG): container finished" podID="493288a9-105f-4ac3-8f6e-2fde520d3d19" containerID="4f6b54fda6fd2e611e1267b94b1d33df9e78737c66fdc769a1fb78e46a2fc845" exitCode=0 Nov 22 05:13:54 crc kubenswrapper[4818]: I1122 05:13:54.773032 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"493288a9-105f-4ac3-8f6e-2fde520d3d19","Type":"ContainerDied","Data":"4f6b54fda6fd2e611e1267b94b1d33df9e78737c66fdc769a1fb78e46a2fc845"} Nov 22 05:13:54 crc kubenswrapper[4818]: I1122 05:13:54.773275 4818 scope.go:117] "RemoveContainer" containerID="4f6b54fda6fd2e611e1267b94b1d33df9e78737c66fdc769a1fb78e46a2fc845" Nov 22 05:13:54 crc kubenswrapper[4818]: I1122 05:13:54.775816 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"493288a9-105f-4ac3-8f6e-2fde520d3d19","Type":"ContainerDied","Data":"2d5ce36095c8d050046b26fd9171d0df09f2ad5e5da1bb6dc5b07dbf4b957fca"} Nov 22 05:13:54 crc kubenswrapper[4818]: I1122 05:13:54.778586 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 05:13:54 crc kubenswrapper[4818]: I1122 05:13:54.807615 4818 scope.go:117] "RemoveContainer" containerID="4f6b54fda6fd2e611e1267b94b1d33df9e78737c66fdc769a1fb78e46a2fc845" Nov 22 05:13:54 crc kubenswrapper[4818]: E1122 05:13:54.808460 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f6b54fda6fd2e611e1267b94b1d33df9e78737c66fdc769a1fb78e46a2fc845\": container with ID starting with 4f6b54fda6fd2e611e1267b94b1d33df9e78737c66fdc769a1fb78e46a2fc845 not found: ID does not exist" containerID="4f6b54fda6fd2e611e1267b94b1d33df9e78737c66fdc769a1fb78e46a2fc845" Nov 22 05:13:54 crc kubenswrapper[4818]: I1122 05:13:54.808510 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f6b54fda6fd2e611e1267b94b1d33df9e78737c66fdc769a1fb78e46a2fc845"} err="failed to get container status \"4f6b54fda6fd2e611e1267b94b1d33df9e78737c66fdc769a1fb78e46a2fc845\": rpc error: code = NotFound desc = could not find container \"4f6b54fda6fd2e611e1267b94b1d33df9e78737c66fdc769a1fb78e46a2fc845\": container with ID starting with 4f6b54fda6fd2e611e1267b94b1d33df9e78737c66fdc769a1fb78e46a2fc845 not found: ID does not exist" Nov 22 05:13:54 crc kubenswrapper[4818]: I1122 05:13:54.878280 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493288a9-105f-4ac3-8f6e-2fde520d3d19-config-data\") pod \"493288a9-105f-4ac3-8f6e-2fde520d3d19\" (UID: \"493288a9-105f-4ac3-8f6e-2fde520d3d19\") " Nov 22 05:13:54 crc kubenswrapper[4818]: I1122 05:13:54.878486 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493288a9-105f-4ac3-8f6e-2fde520d3d19-combined-ca-bundle\") pod \"493288a9-105f-4ac3-8f6e-2fde520d3d19\" (UID: \"493288a9-105f-4ac3-8f6e-2fde520d3d19\") " Nov 22 05:13:54 crc kubenswrapper[4818]: I1122 05:13:54.878553 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bltmn\" (UniqueName: \"kubernetes.io/projected/493288a9-105f-4ac3-8f6e-2fde520d3d19-kube-api-access-bltmn\") pod \"493288a9-105f-4ac3-8f6e-2fde520d3d19\" (UID: \"493288a9-105f-4ac3-8f6e-2fde520d3d19\") " Nov 22 05:13:54 crc kubenswrapper[4818]: I1122 05:13:54.891381 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/493288a9-105f-4ac3-8f6e-2fde520d3d19-kube-api-access-bltmn" (OuterVolumeSpecName: "kube-api-access-bltmn") pod "493288a9-105f-4ac3-8f6e-2fde520d3d19" (UID: "493288a9-105f-4ac3-8f6e-2fde520d3d19"). InnerVolumeSpecName "kube-api-access-bltmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:13:54 crc kubenswrapper[4818]: I1122 05:13:54.908691 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493288a9-105f-4ac3-8f6e-2fde520d3d19-config-data" (OuterVolumeSpecName: "config-data") pod "493288a9-105f-4ac3-8f6e-2fde520d3d19" (UID: "493288a9-105f-4ac3-8f6e-2fde520d3d19"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:54 crc kubenswrapper[4818]: I1122 05:13:54.910127 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493288a9-105f-4ac3-8f6e-2fde520d3d19-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "493288a9-105f-4ac3-8f6e-2fde520d3d19" (UID: "493288a9-105f-4ac3-8f6e-2fde520d3d19"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:54 crc kubenswrapper[4818]: I1122 05:13:54.981058 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493288a9-105f-4ac3-8f6e-2fde520d3d19-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:54 crc kubenswrapper[4818]: I1122 05:13:54.981126 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bltmn\" (UniqueName: \"kubernetes.io/projected/493288a9-105f-4ac3-8f6e-2fde520d3d19-kube-api-access-bltmn\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:54 crc kubenswrapper[4818]: I1122 05:13:54.981137 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493288a9-105f-4ac3-8f6e-2fde520d3d19-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.161122 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.174604 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.182349 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 05:13:55 crc kubenswrapper[4818]: E1122 05:13:55.182706 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdfdf908-61b8-4dcd-9898-fafdee790ca0" containerName="nova-manage" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.182724 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdfdf908-61b8-4dcd-9898-fafdee790ca0" containerName="nova-manage" Nov 22 05:13:55 crc kubenswrapper[4818]: E1122 05:13:55.182737 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34665b87-c9bb-4b11-9406-57bc79ed2e0c" containerName="init" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.182744 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="34665b87-c9bb-4b11-9406-57bc79ed2e0c" containerName="init" Nov 22 05:13:55 crc kubenswrapper[4818]: E1122 05:13:55.182759 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="493288a9-105f-4ac3-8f6e-2fde520d3d19" containerName="nova-scheduler-scheduler" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.182766 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="493288a9-105f-4ac3-8f6e-2fde520d3d19" containerName="nova-scheduler-scheduler" Nov 22 05:13:55 crc kubenswrapper[4818]: E1122 05:13:55.182781 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34665b87-c9bb-4b11-9406-57bc79ed2e0c" containerName="dnsmasq-dns" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.182787 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="34665b87-c9bb-4b11-9406-57bc79ed2e0c" containerName="dnsmasq-dns" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.182945 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdfdf908-61b8-4dcd-9898-fafdee790ca0" containerName="nova-manage" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.182959 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="34665b87-c9bb-4b11-9406-57bc79ed2e0c" containerName="dnsmasq-dns" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.182981 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="493288a9-105f-4ac3-8f6e-2fde520d3d19" containerName="nova-scheduler-scheduler" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.183658 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.189183 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.190950 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.290567 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m629z\" (UniqueName: \"kubernetes.io/projected/f9271e30-e243-4016-bc57-2a10e514deb0-kube-api-access-m629z\") pod \"nova-scheduler-0\" (UID: \"f9271e30-e243-4016-bc57-2a10e514deb0\") " pod="openstack/nova-scheduler-0" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.290619 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9271e30-e243-4016-bc57-2a10e514deb0-config-data\") pod \"nova-scheduler-0\" (UID: \"f9271e30-e243-4016-bc57-2a10e514deb0\") " pod="openstack/nova-scheduler-0" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.290732 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9271e30-e243-4016-bc57-2a10e514deb0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f9271e30-e243-4016-bc57-2a10e514deb0\") " pod="openstack/nova-scheduler-0" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.392311 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m629z\" (UniqueName: \"kubernetes.io/projected/f9271e30-e243-4016-bc57-2a10e514deb0-kube-api-access-m629z\") pod \"nova-scheduler-0\" (UID: \"f9271e30-e243-4016-bc57-2a10e514deb0\") " pod="openstack/nova-scheduler-0" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.392361 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9271e30-e243-4016-bc57-2a10e514deb0-config-data\") pod \"nova-scheduler-0\" (UID: \"f9271e30-e243-4016-bc57-2a10e514deb0\") " pod="openstack/nova-scheduler-0" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.392538 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9271e30-e243-4016-bc57-2a10e514deb0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f9271e30-e243-4016-bc57-2a10e514deb0\") " pod="openstack/nova-scheduler-0" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.398272 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9271e30-e243-4016-bc57-2a10e514deb0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f9271e30-e243-4016-bc57-2a10e514deb0\") " pod="openstack/nova-scheduler-0" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.409501 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m629z\" (UniqueName: \"kubernetes.io/projected/f9271e30-e243-4016-bc57-2a10e514deb0-kube-api-access-m629z\") pod \"nova-scheduler-0\" (UID: \"f9271e30-e243-4016-bc57-2a10e514deb0\") " pod="openstack/nova-scheduler-0" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.411334 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9271e30-e243-4016-bc57-2a10e514deb0-config-data\") pod \"nova-scheduler-0\" (UID: \"f9271e30-e243-4016-bc57-2a10e514deb0\") " pod="openstack/nova-scheduler-0" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.506681 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 05:13:55 crc kubenswrapper[4818]: I1122 05:13:55.947455 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 05:13:55 crc kubenswrapper[4818]: W1122 05:13:55.953218 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9271e30_e243_4016_bc57_2a10e514deb0.slice/crio-3c5851eff2c7dce06536bc664e3dd9c00d67aae988a88640e825951ca232cacb WatchSource:0}: Error finding container 3c5851eff2c7dce06536bc664e3dd9c00d67aae988a88640e825951ca232cacb: Status 404 returned error can't find the container with id 3c5851eff2c7dce06536bc664e3dd9c00d67aae988a88640e825951ca232cacb Nov 22 05:13:56 crc kubenswrapper[4818]: I1122 05:13:56.306385 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="493288a9-105f-4ac3-8f6e-2fde520d3d19" path="/var/lib/kubelet/pods/493288a9-105f-4ac3-8f6e-2fde520d3d19/volumes" Nov 22 05:13:56 crc kubenswrapper[4818]: I1122 05:13:56.795380 4818 generic.go:334] "Generic (PLEG): container finished" podID="7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12" containerID="51dd0c67d382bf1b8bbe35e6cac4fe11396139cdb92c92f82e2aeab8dcf57231" exitCode=0 Nov 22 05:13:56 crc kubenswrapper[4818]: I1122 05:13:56.795628 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12","Type":"ContainerDied","Data":"51dd0c67d382bf1b8bbe35e6cac4fe11396139cdb92c92f82e2aeab8dcf57231"} Nov 22 05:13:56 crc kubenswrapper[4818]: I1122 05:13:56.795937 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12","Type":"ContainerDied","Data":"ee952d2e76d4047b2d0404e9f0bc056a3707dc8795bbe5ad8b5d3876a0d4e6dd"} Nov 22 05:13:56 crc kubenswrapper[4818]: I1122 05:13:56.795960 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee952d2e76d4047b2d0404e9f0bc056a3707dc8795bbe5ad8b5d3876a0d4e6dd" Nov 22 05:13:56 crc kubenswrapper[4818]: I1122 05:13:56.798776 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f9271e30-e243-4016-bc57-2a10e514deb0","Type":"ContainerStarted","Data":"d721bac1efc72e8ca0568e69ec230f2e6b017436544e14e06e30cb74d2b9b7d4"} Nov 22 05:13:56 crc kubenswrapper[4818]: I1122 05:13:56.798853 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f9271e30-e243-4016-bc57-2a10e514deb0","Type":"ContainerStarted","Data":"3c5851eff2c7dce06536bc664e3dd9c00d67aae988a88640e825951ca232cacb"} Nov 22 05:13:56 crc kubenswrapper[4818]: I1122 05:13:56.807551 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 05:13:56 crc kubenswrapper[4818]: I1122 05:13:56.830740 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.830723417 podStartE2EDuration="1.830723417s" podCreationTimestamp="2025-11-22 05:13:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:13:56.82367242 +0000 UTC m=+1589.398088947" watchObservedRunningTime="2025-11-22 05:13:56.830723417 +0000 UTC m=+1589.405139944" Nov 22 05:13:56 crc kubenswrapper[4818]: I1122 05:13:56.929374 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-config-data\") pod \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\" (UID: \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\") " Nov 22 05:13:56 crc kubenswrapper[4818]: I1122 05:13:56.929863 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggwtr\" (UniqueName: \"kubernetes.io/projected/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-kube-api-access-ggwtr\") pod \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\" (UID: \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\") " Nov 22 05:13:56 crc kubenswrapper[4818]: I1122 05:13:56.929954 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-combined-ca-bundle\") pod \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\" (UID: \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\") " Nov 22 05:13:56 crc kubenswrapper[4818]: I1122 05:13:56.930022 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-nova-metadata-tls-certs\") pod \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\" (UID: \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\") " Nov 22 05:13:56 crc kubenswrapper[4818]: I1122 05:13:56.930096 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-logs\") pod \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\" (UID: \"7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12\") " Nov 22 05:13:56 crc kubenswrapper[4818]: I1122 05:13:56.930740 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-logs" (OuterVolumeSpecName: "logs") pod "7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12" (UID: "7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:13:56 crc kubenswrapper[4818]: I1122 05:13:56.932161 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-logs\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:56 crc kubenswrapper[4818]: I1122 05:13:56.935116 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-kube-api-access-ggwtr" (OuterVolumeSpecName: "kube-api-access-ggwtr") pod "7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12" (UID: "7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12"). InnerVolumeSpecName "kube-api-access-ggwtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:13:56 crc kubenswrapper[4818]: I1122 05:13:56.967439 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-config-data" (OuterVolumeSpecName: "config-data") pod "7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12" (UID: "7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:56 crc kubenswrapper[4818]: I1122 05:13:56.979080 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12" (UID: "7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:57 crc kubenswrapper[4818]: I1122 05:13:57.005096 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12" (UID: "7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:57 crc kubenswrapper[4818]: I1122 05:13:57.033822 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggwtr\" (UniqueName: \"kubernetes.io/projected/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-kube-api-access-ggwtr\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:57 crc kubenswrapper[4818]: I1122 05:13:57.033852 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:57 crc kubenswrapper[4818]: I1122 05:13:57.033863 4818 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:57 crc kubenswrapper[4818]: I1122 05:13:57.033872 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:57 crc kubenswrapper[4818]: I1122 05:13:57.812334 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 05:13:57 crc kubenswrapper[4818]: I1122 05:13:57.882795 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 05:13:57 crc kubenswrapper[4818]: I1122 05:13:57.899478 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 05:13:57 crc kubenswrapper[4818]: I1122 05:13:57.906363 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 22 05:13:57 crc kubenswrapper[4818]: E1122 05:13:57.906752 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12" containerName="nova-metadata-log" Nov 22 05:13:57 crc kubenswrapper[4818]: I1122 05:13:57.906771 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12" containerName="nova-metadata-log" Nov 22 05:13:57 crc kubenswrapper[4818]: E1122 05:13:57.906800 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12" containerName="nova-metadata-metadata" Nov 22 05:13:57 crc kubenswrapper[4818]: I1122 05:13:57.906806 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12" containerName="nova-metadata-metadata" Nov 22 05:13:57 crc kubenswrapper[4818]: I1122 05:13:57.906962 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12" containerName="nova-metadata-metadata" Nov 22 05:13:57 crc kubenswrapper[4818]: I1122 05:13:57.906984 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12" containerName="nova-metadata-log" Nov 22 05:13:57 crc kubenswrapper[4818]: I1122 05:13:57.907955 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 05:13:57 crc kubenswrapper[4818]: I1122 05:13:57.912168 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 22 05:13:57 crc kubenswrapper[4818]: I1122 05:13:57.912430 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 22 05:13:57 crc kubenswrapper[4818]: I1122 05:13:57.924765 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 05:13:58 crc kubenswrapper[4818]: I1122 05:13:58.059731 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2b602e6-0bbf-4137-a806-d974b00003d7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e2b602e6-0bbf-4137-a806-d974b00003d7\") " pod="openstack/nova-metadata-0" Nov 22 05:13:58 crc kubenswrapper[4818]: I1122 05:13:58.059918 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2b602e6-0bbf-4137-a806-d974b00003d7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e2b602e6-0bbf-4137-a806-d974b00003d7\") " pod="openstack/nova-metadata-0" Nov 22 05:13:58 crc kubenswrapper[4818]: I1122 05:13:58.059951 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2b602e6-0bbf-4137-a806-d974b00003d7-config-data\") pod \"nova-metadata-0\" (UID: \"e2b602e6-0bbf-4137-a806-d974b00003d7\") " pod="openstack/nova-metadata-0" Nov 22 05:13:58 crc kubenswrapper[4818]: I1122 05:13:58.059982 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc65m\" (UniqueName: \"kubernetes.io/projected/e2b602e6-0bbf-4137-a806-d974b00003d7-kube-api-access-jc65m\") pod \"nova-metadata-0\" (UID: \"e2b602e6-0bbf-4137-a806-d974b00003d7\") " pod="openstack/nova-metadata-0" Nov 22 05:13:58 crc kubenswrapper[4818]: I1122 05:13:58.060008 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2b602e6-0bbf-4137-a806-d974b00003d7-logs\") pod \"nova-metadata-0\" (UID: \"e2b602e6-0bbf-4137-a806-d974b00003d7\") " pod="openstack/nova-metadata-0" Nov 22 05:13:58 crc kubenswrapper[4818]: I1122 05:13:58.163095 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2b602e6-0bbf-4137-a806-d974b00003d7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e2b602e6-0bbf-4137-a806-d974b00003d7\") " pod="openstack/nova-metadata-0" Nov 22 05:13:58 crc kubenswrapper[4818]: I1122 05:13:58.163334 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2b602e6-0bbf-4137-a806-d974b00003d7-config-data\") pod \"nova-metadata-0\" (UID: \"e2b602e6-0bbf-4137-a806-d974b00003d7\") " pod="openstack/nova-metadata-0" Nov 22 05:13:58 crc kubenswrapper[4818]: I1122 05:13:58.163413 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc65m\" (UniqueName: \"kubernetes.io/projected/e2b602e6-0bbf-4137-a806-d974b00003d7-kube-api-access-jc65m\") pod \"nova-metadata-0\" (UID: \"e2b602e6-0bbf-4137-a806-d974b00003d7\") " pod="openstack/nova-metadata-0" Nov 22 05:13:58 crc kubenswrapper[4818]: I1122 05:13:58.163478 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2b602e6-0bbf-4137-a806-d974b00003d7-logs\") pod \"nova-metadata-0\" (UID: \"e2b602e6-0bbf-4137-a806-d974b00003d7\") " pod="openstack/nova-metadata-0" Nov 22 05:13:58 crc kubenswrapper[4818]: I1122 05:13:58.163622 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2b602e6-0bbf-4137-a806-d974b00003d7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e2b602e6-0bbf-4137-a806-d974b00003d7\") " pod="openstack/nova-metadata-0" Nov 22 05:13:58 crc kubenswrapper[4818]: I1122 05:13:58.164822 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2b602e6-0bbf-4137-a806-d974b00003d7-logs\") pod \"nova-metadata-0\" (UID: \"e2b602e6-0bbf-4137-a806-d974b00003d7\") " pod="openstack/nova-metadata-0" Nov 22 05:13:58 crc kubenswrapper[4818]: I1122 05:13:58.168240 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2b602e6-0bbf-4137-a806-d974b00003d7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e2b602e6-0bbf-4137-a806-d974b00003d7\") " pod="openstack/nova-metadata-0" Nov 22 05:13:58 crc kubenswrapper[4818]: I1122 05:13:58.168969 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2b602e6-0bbf-4137-a806-d974b00003d7-config-data\") pod \"nova-metadata-0\" (UID: \"e2b602e6-0bbf-4137-a806-d974b00003d7\") " pod="openstack/nova-metadata-0" Nov 22 05:13:58 crc kubenswrapper[4818]: I1122 05:13:58.169099 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2b602e6-0bbf-4137-a806-d974b00003d7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e2b602e6-0bbf-4137-a806-d974b00003d7\") " pod="openstack/nova-metadata-0" Nov 22 05:13:58 crc kubenswrapper[4818]: I1122 05:13:58.179479 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc65m\" (UniqueName: \"kubernetes.io/projected/e2b602e6-0bbf-4137-a806-d974b00003d7-kube-api-access-jc65m\") pod \"nova-metadata-0\" (UID: \"e2b602e6-0bbf-4137-a806-d974b00003d7\") " pod="openstack/nova-metadata-0" Nov 22 05:13:58 crc kubenswrapper[4818]: I1122 05:13:58.228841 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 05:13:58 crc kubenswrapper[4818]: I1122 05:13:58.303427 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12" path="/var/lib/kubelet/pods/7a6786f7-e9e0-4bcc-9115-1bfbbf7e8b12/volumes" Nov 22 05:13:58 crc kubenswrapper[4818]: I1122 05:13:58.686875 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 05:13:58 crc kubenswrapper[4818]: I1122 05:13:58.826698 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e2b602e6-0bbf-4137-a806-d974b00003d7","Type":"ContainerStarted","Data":"957b25d11005df5c1a56bcdd2920e462a60151bc10ca98b0aa70cf1e529d9ca1"} Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.521896 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.598196 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-internal-tls-certs\") pod \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.598354 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-config-data\") pod \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.598429 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-combined-ca-bundle\") pod \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.598515 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flnwx\" (UniqueName: \"kubernetes.io/projected/cb13e5c9-26fd-44d4-b745-b77ce47582ac-kube-api-access-flnwx\") pod \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.598597 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-public-tls-certs\") pod \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.598665 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb13e5c9-26fd-44d4-b745-b77ce47582ac-logs\") pod \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\" (UID: \"cb13e5c9-26fd-44d4-b745-b77ce47582ac\") " Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.599192 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb13e5c9-26fd-44d4-b745-b77ce47582ac-logs" (OuterVolumeSpecName: "logs") pod "cb13e5c9-26fd-44d4-b745-b77ce47582ac" (UID: "cb13e5c9-26fd-44d4-b745-b77ce47582ac"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.599501 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb13e5c9-26fd-44d4-b745-b77ce47582ac-logs\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.603356 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb13e5c9-26fd-44d4-b745-b77ce47582ac-kube-api-access-flnwx" (OuterVolumeSpecName: "kube-api-access-flnwx") pod "cb13e5c9-26fd-44d4-b745-b77ce47582ac" (UID: "cb13e5c9-26fd-44d4-b745-b77ce47582ac"). InnerVolumeSpecName "kube-api-access-flnwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.625999 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb13e5c9-26fd-44d4-b745-b77ce47582ac" (UID: "cb13e5c9-26fd-44d4-b745-b77ce47582ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.627386 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-config-data" (OuterVolumeSpecName: "config-data") pod "cb13e5c9-26fd-44d4-b745-b77ce47582ac" (UID: "cb13e5c9-26fd-44d4-b745-b77ce47582ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.643926 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "cb13e5c9-26fd-44d4-b745-b77ce47582ac" (UID: "cb13e5c9-26fd-44d4-b745-b77ce47582ac"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.651128 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "cb13e5c9-26fd-44d4-b745-b77ce47582ac" (UID: "cb13e5c9-26fd-44d4-b745-b77ce47582ac"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.727750 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.727781 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.727792 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flnwx\" (UniqueName: \"kubernetes.io/projected/cb13e5c9-26fd-44d4-b745-b77ce47582ac-kube-api-access-flnwx\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.727802 4818 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.727810 4818 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb13e5c9-26fd-44d4-b745-b77ce47582ac-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.837932 4818 generic.go:334] "Generic (PLEG): container finished" podID="cb13e5c9-26fd-44d4-b745-b77ce47582ac" containerID="179c2ad91ce19d61606d928971b0be2664258d48a8a4b62a733db9e969aa162a" exitCode=0 Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.838011 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.838028 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cb13e5c9-26fd-44d4-b745-b77ce47582ac","Type":"ContainerDied","Data":"179c2ad91ce19d61606d928971b0be2664258d48a8a4b62a733db9e969aa162a"} Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.838068 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cb13e5c9-26fd-44d4-b745-b77ce47582ac","Type":"ContainerDied","Data":"965c1341308b1275f5207a87cd376784fbc2da6264db1075c39dad63a60073f8"} Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.838086 4818 scope.go:117] "RemoveContainer" containerID="179c2ad91ce19d61606d928971b0be2664258d48a8a4b62a733db9e969aa162a" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.841619 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e2b602e6-0bbf-4137-a806-d974b00003d7","Type":"ContainerStarted","Data":"dc4ab34e2a6fd236787ae049e14146c9128cb3f27344508936291699d19af450"} Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.841671 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e2b602e6-0bbf-4137-a806-d974b00003d7","Type":"ContainerStarted","Data":"715f4717d2e5907fd50d535f4b7514284ee727a45f666249c9bcb2e468addb53"} Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.869429 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.869406101 podStartE2EDuration="2.869406101s" podCreationTimestamp="2025-11-22 05:13:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:13:59.86562245 +0000 UTC m=+1592.440039007" watchObservedRunningTime="2025-11-22 05:13:59.869406101 +0000 UTC m=+1592.443822638" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.882728 4818 scope.go:117] "RemoveContainer" containerID="d43de0a0e44af675cab5f29fab0bfaae39db5c71d9d59bccd5b23fa8bb39aac3" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.901924 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.919665 4818 scope.go:117] "RemoveContainer" containerID="179c2ad91ce19d61606d928971b0be2664258d48a8a4b62a733db9e969aa162a" Nov 22 05:13:59 crc kubenswrapper[4818]: E1122 05:13:59.920150 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"179c2ad91ce19d61606d928971b0be2664258d48a8a4b62a733db9e969aa162a\": container with ID starting with 179c2ad91ce19d61606d928971b0be2664258d48a8a4b62a733db9e969aa162a not found: ID does not exist" containerID="179c2ad91ce19d61606d928971b0be2664258d48a8a4b62a733db9e969aa162a" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.920186 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"179c2ad91ce19d61606d928971b0be2664258d48a8a4b62a733db9e969aa162a"} err="failed to get container status \"179c2ad91ce19d61606d928971b0be2664258d48a8a4b62a733db9e969aa162a\": rpc error: code = NotFound desc = could not find container \"179c2ad91ce19d61606d928971b0be2664258d48a8a4b62a733db9e969aa162a\": container with ID starting with 179c2ad91ce19d61606d928971b0be2664258d48a8a4b62a733db9e969aa162a not found: ID does not exist" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.920208 4818 scope.go:117] "RemoveContainer" containerID="d43de0a0e44af675cab5f29fab0bfaae39db5c71d9d59bccd5b23fa8bb39aac3" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.920396 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 22 05:13:59 crc kubenswrapper[4818]: E1122 05:13:59.920454 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d43de0a0e44af675cab5f29fab0bfaae39db5c71d9d59bccd5b23fa8bb39aac3\": container with ID starting with d43de0a0e44af675cab5f29fab0bfaae39db5c71d9d59bccd5b23fa8bb39aac3 not found: ID does not exist" containerID="d43de0a0e44af675cab5f29fab0bfaae39db5c71d9d59bccd5b23fa8bb39aac3" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.920477 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d43de0a0e44af675cab5f29fab0bfaae39db5c71d9d59bccd5b23fa8bb39aac3"} err="failed to get container status \"d43de0a0e44af675cab5f29fab0bfaae39db5c71d9d59bccd5b23fa8bb39aac3\": rpc error: code = NotFound desc = could not find container \"d43de0a0e44af675cab5f29fab0bfaae39db5c71d9d59bccd5b23fa8bb39aac3\": container with ID starting with d43de0a0e44af675cab5f29fab0bfaae39db5c71d9d59bccd5b23fa8bb39aac3 not found: ID does not exist" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.932972 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 22 05:13:59 crc kubenswrapper[4818]: E1122 05:13:59.940131 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb13e5c9-26fd-44d4-b745-b77ce47582ac" containerName="nova-api-api" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.940161 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb13e5c9-26fd-44d4-b745-b77ce47582ac" containerName="nova-api-api" Nov 22 05:13:59 crc kubenswrapper[4818]: E1122 05:13:59.940220 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb13e5c9-26fd-44d4-b745-b77ce47582ac" containerName="nova-api-log" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.940228 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb13e5c9-26fd-44d4-b745-b77ce47582ac" containerName="nova-api-log" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.940563 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb13e5c9-26fd-44d4-b745-b77ce47582ac" containerName="nova-api-api" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.940631 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb13e5c9-26fd-44d4-b745-b77ce47582ac" containerName="nova-api-log" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.942049 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.943157 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.945374 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.945536 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 22 05:13:59 crc kubenswrapper[4818]: I1122 05:13:59.945650 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.033270 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed42de57-a99e-4940-9874-49f6f703f8aa-config-data\") pod \"nova-api-0\" (UID: \"ed42de57-a99e-4940-9874-49f6f703f8aa\") " pod="openstack/nova-api-0" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.033336 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed42de57-a99e-4940-9874-49f6f703f8aa-public-tls-certs\") pod \"nova-api-0\" (UID: \"ed42de57-a99e-4940-9874-49f6f703f8aa\") " pod="openstack/nova-api-0" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.033376 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed42de57-a99e-4940-9874-49f6f703f8aa-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ed42de57-a99e-4940-9874-49f6f703f8aa\") " pod="openstack/nova-api-0" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.033398 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scjnn\" (UniqueName: \"kubernetes.io/projected/ed42de57-a99e-4940-9874-49f6f703f8aa-kube-api-access-scjnn\") pod \"nova-api-0\" (UID: \"ed42de57-a99e-4940-9874-49f6f703f8aa\") " pod="openstack/nova-api-0" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.033420 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed42de57-a99e-4940-9874-49f6f703f8aa-logs\") pod \"nova-api-0\" (UID: \"ed42de57-a99e-4940-9874-49f6f703f8aa\") " pod="openstack/nova-api-0" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.033740 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed42de57-a99e-4940-9874-49f6f703f8aa-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ed42de57-a99e-4940-9874-49f6f703f8aa\") " pod="openstack/nova-api-0" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.136083 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed42de57-a99e-4940-9874-49f6f703f8aa-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ed42de57-a99e-4940-9874-49f6f703f8aa\") " pod="openstack/nova-api-0" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.136189 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed42de57-a99e-4940-9874-49f6f703f8aa-config-data\") pod \"nova-api-0\" (UID: \"ed42de57-a99e-4940-9874-49f6f703f8aa\") " pod="openstack/nova-api-0" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.136209 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed42de57-a99e-4940-9874-49f6f703f8aa-public-tls-certs\") pod \"nova-api-0\" (UID: \"ed42de57-a99e-4940-9874-49f6f703f8aa\") " pod="openstack/nova-api-0" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.136235 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed42de57-a99e-4940-9874-49f6f703f8aa-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ed42de57-a99e-4940-9874-49f6f703f8aa\") " pod="openstack/nova-api-0" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.136268 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scjnn\" (UniqueName: \"kubernetes.io/projected/ed42de57-a99e-4940-9874-49f6f703f8aa-kube-api-access-scjnn\") pod \"nova-api-0\" (UID: \"ed42de57-a99e-4940-9874-49f6f703f8aa\") " pod="openstack/nova-api-0" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.136290 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed42de57-a99e-4940-9874-49f6f703f8aa-logs\") pod \"nova-api-0\" (UID: \"ed42de57-a99e-4940-9874-49f6f703f8aa\") " pod="openstack/nova-api-0" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.136985 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed42de57-a99e-4940-9874-49f6f703f8aa-logs\") pod \"nova-api-0\" (UID: \"ed42de57-a99e-4940-9874-49f6f703f8aa\") " pod="openstack/nova-api-0" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.141305 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed42de57-a99e-4940-9874-49f6f703f8aa-config-data\") pod \"nova-api-0\" (UID: \"ed42de57-a99e-4940-9874-49f6f703f8aa\") " pod="openstack/nova-api-0" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.142718 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed42de57-a99e-4940-9874-49f6f703f8aa-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ed42de57-a99e-4940-9874-49f6f703f8aa\") " pod="openstack/nova-api-0" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.142773 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed42de57-a99e-4940-9874-49f6f703f8aa-public-tls-certs\") pod \"nova-api-0\" (UID: \"ed42de57-a99e-4940-9874-49f6f703f8aa\") " pod="openstack/nova-api-0" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.143273 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed42de57-a99e-4940-9874-49f6f703f8aa-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ed42de57-a99e-4940-9874-49f6f703f8aa\") " pod="openstack/nova-api-0" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.155285 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scjnn\" (UniqueName: \"kubernetes.io/projected/ed42de57-a99e-4940-9874-49f6f703f8aa-kube-api-access-scjnn\") pod \"nova-api-0\" (UID: \"ed42de57-a99e-4940-9874-49f6f703f8aa\") " pod="openstack/nova-api-0" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.259397 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.309800 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb13e5c9-26fd-44d4-b745-b77ce47582ac" path="/var/lib/kubelet/pods/cb13e5c9-26fd-44d4-b745-b77ce47582ac/volumes" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.507042 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.745015 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 05:14:00 crc kubenswrapper[4818]: W1122 05:14:00.750284 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded42de57_a99e_4940_9874_49f6f703f8aa.slice/crio-afcc4d67fa7a59360fd4d14bf88a5ff7c3fa41e44dba07298b9947fb6caeaaa4 WatchSource:0}: Error finding container afcc4d67fa7a59360fd4d14bf88a5ff7c3fa41e44dba07298b9947fb6caeaaa4: Status 404 returned error can't find the container with id afcc4d67fa7a59360fd4d14bf88a5ff7c3fa41e44dba07298b9947fb6caeaaa4 Nov 22 05:14:00 crc kubenswrapper[4818]: I1122 05:14:00.855351 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ed42de57-a99e-4940-9874-49f6f703f8aa","Type":"ContainerStarted","Data":"afcc4d67fa7a59360fd4d14bf88a5ff7c3fa41e44dba07298b9947fb6caeaaa4"} Nov 22 05:14:01 crc kubenswrapper[4818]: I1122 05:14:01.865939 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ed42de57-a99e-4940-9874-49f6f703f8aa","Type":"ContainerStarted","Data":"2075b6ac4815fa0d7a11773f283885db38d4b116ffff396e58caedb31839d18b"} Nov 22 05:14:01 crc kubenswrapper[4818]: I1122 05:14:01.866592 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ed42de57-a99e-4940-9874-49f6f703f8aa","Type":"ContainerStarted","Data":"4f3a3497c3058a2fd4858c672e84bdbdd211e2df13ea2d3bc5fadbd2b8e931a3"} Nov 22 05:14:01 crc kubenswrapper[4818]: I1122 05:14:01.908525 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.90849284 podStartE2EDuration="2.90849284s" podCreationTimestamp="2025-11-22 05:13:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:14:01.888675094 +0000 UTC m=+1594.463091621" watchObservedRunningTime="2025-11-22 05:14:01.90849284 +0000 UTC m=+1594.482909377" Nov 22 05:14:03 crc kubenswrapper[4818]: I1122 05:14:03.230067 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 22 05:14:03 crc kubenswrapper[4818]: I1122 05:14:03.230157 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 22 05:14:05 crc kubenswrapper[4818]: I1122 05:14:05.507191 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 22 05:14:05 crc kubenswrapper[4818]: I1122 05:14:05.535711 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 22 05:14:05 crc kubenswrapper[4818]: I1122 05:14:05.948383 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 22 05:14:08 crc kubenswrapper[4818]: I1122 05:14:08.229655 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 22 05:14:08 crc kubenswrapper[4818]: I1122 05:14:08.229927 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 22 05:14:09 crc kubenswrapper[4818]: I1122 05:14:09.250503 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e2b602e6-0bbf-4137-a806-d974b00003d7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.186:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 22 05:14:09 crc kubenswrapper[4818]: I1122 05:14:09.250530 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e2b602e6-0bbf-4137-a806-d974b00003d7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.186:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 22 05:14:10 crc kubenswrapper[4818]: I1122 05:14:10.260627 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 05:14:10 crc kubenswrapper[4818]: I1122 05:14:10.260705 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 05:14:11 crc kubenswrapper[4818]: I1122 05:14:11.282609 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ed42de57-a99e-4940-9874-49f6f703f8aa" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.187:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 05:14:11 crc kubenswrapper[4818]: I1122 05:14:11.282602 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ed42de57-a99e-4940-9874-49f6f703f8aa" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.187:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 22 05:14:16 crc kubenswrapper[4818]: I1122 05:14:16.085680 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 22 05:14:18 crc kubenswrapper[4818]: I1122 05:14:18.237097 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 22 05:14:18 crc kubenswrapper[4818]: I1122 05:14:18.238198 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 22 05:14:18 crc kubenswrapper[4818]: I1122 05:14:18.243365 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 22 05:14:19 crc kubenswrapper[4818]: I1122 05:14:19.061301 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 22 05:14:20 crc kubenswrapper[4818]: I1122 05:14:20.268184 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 22 05:14:20 crc kubenswrapper[4818]: I1122 05:14:20.268320 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 22 05:14:20 crc kubenswrapper[4818]: I1122 05:14:20.268668 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 22 05:14:20 crc kubenswrapper[4818]: I1122 05:14:20.268737 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 22 05:14:20 crc kubenswrapper[4818]: I1122 05:14:20.274656 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 22 05:14:20 crc kubenswrapper[4818]: I1122 05:14:20.277447 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 22 05:14:21 crc kubenswrapper[4818]: I1122 05:14:21.264684 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:14:21 crc kubenswrapper[4818]: I1122 05:14:21.265038 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:14:29 crc kubenswrapper[4818]: I1122 05:14:29.063588 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 05:14:30 crc kubenswrapper[4818]: I1122 05:14:30.240679 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 05:14:33 crc kubenswrapper[4818]: I1122 05:14:33.213620 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="69f72526-235a-4079-a089-43e1b93aef9f" containerName="rabbitmq" containerID="cri-o://c51c9537d0a14d91d8b78ef778dfb2df074295aaa1b7aea71a8070be82aeb023" gracePeriod=604796 Nov 22 05:14:34 crc kubenswrapper[4818]: I1122 05:14:34.222331 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="67e09220-fd61-4785-9de7-90b3ffaf0157" containerName="rabbitmq" containerID="cri-o://2011cd97d070b8a754b3acec1fbdf7b22170264b4f513d61217568bcbbd91c82" gracePeriod=604797 Nov 22 05:14:36 crc kubenswrapper[4818]: I1122 05:14:36.662900 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="69f72526-235a-4079-a089-43e1b93aef9f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Nov 22 05:14:37 crc kubenswrapper[4818]: I1122 05:14:37.027651 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="67e09220-fd61-4785-9de7-90b3ffaf0157" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.817983 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.953743 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-erlang-cookie\") pod \"69f72526-235a-4079-a089-43e1b93aef9f\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.953811 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/69f72526-235a-4079-a089-43e1b93aef9f-erlang-cookie-secret\") pod \"69f72526-235a-4079-a089-43e1b93aef9f\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.953841 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-plugins\") pod \"69f72526-235a-4079-a089-43e1b93aef9f\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.953890 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/69f72526-235a-4079-a089-43e1b93aef9f-pod-info\") pod \"69f72526-235a-4079-a089-43e1b93aef9f\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.953916 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/69f72526-235a-4079-a089-43e1b93aef9f-config-data\") pod \"69f72526-235a-4079-a089-43e1b93aef9f\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.953941 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/69f72526-235a-4079-a089-43e1b93aef9f-server-conf\") pod \"69f72526-235a-4079-a089-43e1b93aef9f\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.953979 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-tls\") pod \"69f72526-235a-4079-a089-43e1b93aef9f\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.954013 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9dn6\" (UniqueName: \"kubernetes.io/projected/69f72526-235a-4079-a089-43e1b93aef9f-kube-api-access-f9dn6\") pod \"69f72526-235a-4079-a089-43e1b93aef9f\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.954045 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-confd\") pod \"69f72526-235a-4079-a089-43e1b93aef9f\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.954080 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"69f72526-235a-4079-a089-43e1b93aef9f\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.954129 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/69f72526-235a-4079-a089-43e1b93aef9f-plugins-conf\") pod \"69f72526-235a-4079-a089-43e1b93aef9f\" (UID: \"69f72526-235a-4079-a089-43e1b93aef9f\") " Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.954470 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "69f72526-235a-4079-a089-43e1b93aef9f" (UID: "69f72526-235a-4079-a089-43e1b93aef9f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.955488 4818 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.959992 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "69f72526-235a-4079-a089-43e1b93aef9f" (UID: "69f72526-235a-4079-a089-43e1b93aef9f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.961108 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69f72526-235a-4079-a089-43e1b93aef9f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "69f72526-235a-4079-a089-43e1b93aef9f" (UID: "69f72526-235a-4079-a089-43e1b93aef9f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.962836 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "persistence") pod "69f72526-235a-4079-a089-43e1b93aef9f" (UID: "69f72526-235a-4079-a089-43e1b93aef9f"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.964539 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69f72526-235a-4079-a089-43e1b93aef9f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "69f72526-235a-4079-a089-43e1b93aef9f" (UID: "69f72526-235a-4079-a089-43e1b93aef9f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.967549 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/69f72526-235a-4079-a089-43e1b93aef9f-pod-info" (OuterVolumeSpecName: "pod-info") pod "69f72526-235a-4079-a089-43e1b93aef9f" (UID: "69f72526-235a-4079-a089-43e1b93aef9f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.969543 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69f72526-235a-4079-a089-43e1b93aef9f-kube-api-access-f9dn6" (OuterVolumeSpecName: "kube-api-access-f9dn6") pod "69f72526-235a-4079-a089-43e1b93aef9f" (UID: "69f72526-235a-4079-a089-43e1b93aef9f"). InnerVolumeSpecName "kube-api-access-f9dn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.969920 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "69f72526-235a-4079-a089-43e1b93aef9f" (UID: "69f72526-235a-4079-a089-43e1b93aef9f"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:14:39 crc kubenswrapper[4818]: I1122 05:14:39.995660 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69f72526-235a-4079-a089-43e1b93aef9f-config-data" (OuterVolumeSpecName: "config-data") pod "69f72526-235a-4079-a089-43e1b93aef9f" (UID: "69f72526-235a-4079-a089-43e1b93aef9f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.057704 4818 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/69f72526-235a-4079-a089-43e1b93aef9f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.057740 4818 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.057752 4818 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/69f72526-235a-4079-a089-43e1b93aef9f-pod-info\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.057764 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/69f72526-235a-4079-a089-43e1b93aef9f-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.057774 4818 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.057786 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9dn6\" (UniqueName: \"kubernetes.io/projected/69f72526-235a-4079-a089-43e1b93aef9f-kube-api-access-f9dn6\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.057818 4818 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.057831 4818 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/69f72526-235a-4079-a089-43e1b93aef9f-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.076963 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69f72526-235a-4079-a089-43e1b93aef9f-server-conf" (OuterVolumeSpecName: "server-conf") pod "69f72526-235a-4079-a089-43e1b93aef9f" (UID: "69f72526-235a-4079-a089-43e1b93aef9f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.083560 4818 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.099558 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "69f72526-235a-4079-a089-43e1b93aef9f" (UID: "69f72526-235a-4079-a089-43e1b93aef9f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.159727 4818 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/69f72526-235a-4079-a089-43e1b93aef9f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.159776 4818 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.159786 4818 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/69f72526-235a-4079-a089-43e1b93aef9f-server-conf\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.249370 4818 generic.go:334] "Generic (PLEG): container finished" podID="69f72526-235a-4079-a089-43e1b93aef9f" containerID="c51c9537d0a14d91d8b78ef778dfb2df074295aaa1b7aea71a8070be82aeb023" exitCode=0 Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.249413 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"69f72526-235a-4079-a089-43e1b93aef9f","Type":"ContainerDied","Data":"c51c9537d0a14d91d8b78ef778dfb2df074295aaa1b7aea71a8070be82aeb023"} Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.249438 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"69f72526-235a-4079-a089-43e1b93aef9f","Type":"ContainerDied","Data":"a508f74629c62988b53a3316a12fe97415911f0975215fa692d580fc2c3848c8"} Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.249441 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.249455 4818 scope.go:117] "RemoveContainer" containerID="c51c9537d0a14d91d8b78ef778dfb2df074295aaa1b7aea71a8070be82aeb023" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.295191 4818 scope.go:117] "RemoveContainer" containerID="9f4ec022a951e769e06c755c5e3ee29d2cdb4bf72c3e18edaf341e5d425f9888" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.328143 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.335815 4818 scope.go:117] "RemoveContainer" containerID="c51c9537d0a14d91d8b78ef778dfb2df074295aaa1b7aea71a8070be82aeb023" Nov 22 05:14:40 crc kubenswrapper[4818]: E1122 05:14:40.340752 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c51c9537d0a14d91d8b78ef778dfb2df074295aaa1b7aea71a8070be82aeb023\": container with ID starting with c51c9537d0a14d91d8b78ef778dfb2df074295aaa1b7aea71a8070be82aeb023 not found: ID does not exist" containerID="c51c9537d0a14d91d8b78ef778dfb2df074295aaa1b7aea71a8070be82aeb023" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.340824 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c51c9537d0a14d91d8b78ef778dfb2df074295aaa1b7aea71a8070be82aeb023"} err="failed to get container status \"c51c9537d0a14d91d8b78ef778dfb2df074295aaa1b7aea71a8070be82aeb023\": rpc error: code = NotFound desc = could not find container \"c51c9537d0a14d91d8b78ef778dfb2df074295aaa1b7aea71a8070be82aeb023\": container with ID starting with c51c9537d0a14d91d8b78ef778dfb2df074295aaa1b7aea71a8070be82aeb023 not found: ID does not exist" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.340853 4818 scope.go:117] "RemoveContainer" containerID="9f4ec022a951e769e06c755c5e3ee29d2cdb4bf72c3e18edaf341e5d425f9888" Nov 22 05:14:40 crc kubenswrapper[4818]: E1122 05:14:40.341649 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f4ec022a951e769e06c755c5e3ee29d2cdb4bf72c3e18edaf341e5d425f9888\": container with ID starting with 9f4ec022a951e769e06c755c5e3ee29d2cdb4bf72c3e18edaf341e5d425f9888 not found: ID does not exist" containerID="9f4ec022a951e769e06c755c5e3ee29d2cdb4bf72c3e18edaf341e5d425f9888" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.341717 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f4ec022a951e769e06c755c5e3ee29d2cdb4bf72c3e18edaf341e5d425f9888"} err="failed to get container status \"9f4ec022a951e769e06c755c5e3ee29d2cdb4bf72c3e18edaf341e5d425f9888\": rpc error: code = NotFound desc = could not find container \"9f4ec022a951e769e06c755c5e3ee29d2cdb4bf72c3e18edaf341e5d425f9888\": container with ID starting with 9f4ec022a951e769e06c755c5e3ee29d2cdb4bf72c3e18edaf341e5d425f9888 not found: ID does not exist" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.347111 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.357914 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 05:14:40 crc kubenswrapper[4818]: E1122 05:14:40.359117 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69f72526-235a-4079-a089-43e1b93aef9f" containerName="setup-container" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.359139 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="69f72526-235a-4079-a089-43e1b93aef9f" containerName="setup-container" Nov 22 05:14:40 crc kubenswrapper[4818]: E1122 05:14:40.359192 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69f72526-235a-4079-a089-43e1b93aef9f" containerName="rabbitmq" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.359201 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="69f72526-235a-4079-a089-43e1b93aef9f" containerName="rabbitmq" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.359579 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="69f72526-235a-4079-a089-43e1b93aef9f" containerName="rabbitmq" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.364032 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.367693 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.368005 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-26pzg" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.368282 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.368392 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.368491 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.379047 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.383593 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.383912 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.571923 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-server-conf\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.572185 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-config-data\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.572210 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.572237 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-pod-info\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.572274 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.572288 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.572341 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.572390 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.572442 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.572463 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.572530 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfh72\" (UniqueName: \"kubernetes.io/projected/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-kube-api-access-zfh72\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.674542 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfh72\" (UniqueName: \"kubernetes.io/projected/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-kube-api-access-zfh72\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.674629 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-server-conf\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.674704 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-config-data\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.674743 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.674769 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-pod-info\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.674793 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.674807 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.674825 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.674851 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.674868 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.674882 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.675297 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.675509 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.676209 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-server-conf\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.676221 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-config-data\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.676301 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.680829 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.681795 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.683326 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.686461 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.693106 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-pod-info\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.695893 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfh72\" (UniqueName: \"kubernetes.io/projected/70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86-kube-api-access-zfh72\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.724935 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86\") " pod="openstack/rabbitmq-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.782922 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.980333 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/67e09220-fd61-4785-9de7-90b3ffaf0157-pod-info\") pod \"67e09220-fd61-4785-9de7-90b3ffaf0157\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.981989 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/67e09220-fd61-4785-9de7-90b3ffaf0157-config-data\") pod \"67e09220-fd61-4785-9de7-90b3ffaf0157\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.982160 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/67e09220-fd61-4785-9de7-90b3ffaf0157-erlang-cookie-secret\") pod \"67e09220-fd61-4785-9de7-90b3ffaf0157\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.982486 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-confd\") pod \"67e09220-fd61-4785-9de7-90b3ffaf0157\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.982611 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-erlang-cookie\") pod \"67e09220-fd61-4785-9de7-90b3ffaf0157\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.983397 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/67e09220-fd61-4785-9de7-90b3ffaf0157-plugins-conf\") pod \"67e09220-fd61-4785-9de7-90b3ffaf0157\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.983633 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqs9r\" (UniqueName: \"kubernetes.io/projected/67e09220-fd61-4785-9de7-90b3ffaf0157-kube-api-access-bqs9r\") pod \"67e09220-fd61-4785-9de7-90b3ffaf0157\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.983920 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"67e09220-fd61-4785-9de7-90b3ffaf0157\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.983146 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "67e09220-fd61-4785-9de7-90b3ffaf0157" (UID: "67e09220-fd61-4785-9de7-90b3ffaf0157"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.984699 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67e09220-fd61-4785-9de7-90b3ffaf0157-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "67e09220-fd61-4785-9de7-90b3ffaf0157" (UID: "67e09220-fd61-4785-9de7-90b3ffaf0157"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.985580 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/67e09220-fd61-4785-9de7-90b3ffaf0157-pod-info" (OuterVolumeSpecName: "pod-info") pod "67e09220-fd61-4785-9de7-90b3ffaf0157" (UID: "67e09220-fd61-4785-9de7-90b3ffaf0157"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.985890 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-plugins\") pod \"67e09220-fd61-4785-9de7-90b3ffaf0157\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.986052 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/67e09220-fd61-4785-9de7-90b3ffaf0157-server-conf\") pod \"67e09220-fd61-4785-9de7-90b3ffaf0157\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.986087 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "67e09220-fd61-4785-9de7-90b3ffaf0157" (UID: "67e09220-fd61-4785-9de7-90b3ffaf0157"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:14:40 crc kubenswrapper[4818]: I1122 05:14:40.986506 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-tls\") pod \"67e09220-fd61-4785-9de7-90b3ffaf0157\" (UID: \"67e09220-fd61-4785-9de7-90b3ffaf0157\") " Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.003771 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.004562 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67e09220-fd61-4785-9de7-90b3ffaf0157-kube-api-access-bqs9r" (OuterVolumeSpecName: "kube-api-access-bqs9r") pod "67e09220-fd61-4785-9de7-90b3ffaf0157" (UID: "67e09220-fd61-4785-9de7-90b3ffaf0157"). InnerVolumeSpecName "kube-api-access-bqs9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.005295 4818 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.005493 4818 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/67e09220-fd61-4785-9de7-90b3ffaf0157-pod-info\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.005578 4818 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.005674 4818 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/67e09220-fd61-4785-9de7-90b3ffaf0157-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.005892 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqs9r\" (UniqueName: \"kubernetes.io/projected/67e09220-fd61-4785-9de7-90b3ffaf0157-kube-api-access-bqs9r\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.028514 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67e09220-fd61-4785-9de7-90b3ffaf0157-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "67e09220-fd61-4785-9de7-90b3ffaf0157" (UID: "67e09220-fd61-4785-9de7-90b3ffaf0157"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.032050 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "67e09220-fd61-4785-9de7-90b3ffaf0157" (UID: "67e09220-fd61-4785-9de7-90b3ffaf0157"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.053442 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "67e09220-fd61-4785-9de7-90b3ffaf0157" (UID: "67e09220-fd61-4785-9de7-90b3ffaf0157"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.121816 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67e09220-fd61-4785-9de7-90b3ffaf0157-config-data" (OuterVolumeSpecName: "config-data") pod "67e09220-fd61-4785-9de7-90b3ffaf0157" (UID: "67e09220-fd61-4785-9de7-90b3ffaf0157"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.138226 4818 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.138305 4818 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.138317 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/67e09220-fd61-4785-9de7-90b3ffaf0157-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.138327 4818 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/67e09220-fd61-4785-9de7-90b3ffaf0157-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.190190 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "67e09220-fd61-4785-9de7-90b3ffaf0157" (UID: "67e09220-fd61-4785-9de7-90b3ffaf0157"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.219072 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67e09220-fd61-4785-9de7-90b3ffaf0157-server-conf" (OuterVolumeSpecName: "server-conf") pod "67e09220-fd61-4785-9de7-90b3ffaf0157" (UID: "67e09220-fd61-4785-9de7-90b3ffaf0157"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.224651 4818 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.239681 4818 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/67e09220-fd61-4785-9de7-90b3ffaf0157-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.239714 4818 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.239728 4818 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/67e09220-fd61-4785-9de7-90b3ffaf0157-server-conf\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.262897 4818 generic.go:334] "Generic (PLEG): container finished" podID="67e09220-fd61-4785-9de7-90b3ffaf0157" containerID="2011cd97d070b8a754b3acec1fbdf7b22170264b4f513d61217568bcbbd91c82" exitCode=0 Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.262943 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"67e09220-fd61-4785-9de7-90b3ffaf0157","Type":"ContainerDied","Data":"2011cd97d070b8a754b3acec1fbdf7b22170264b4f513d61217568bcbbd91c82"} Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.262981 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.263017 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"67e09220-fd61-4785-9de7-90b3ffaf0157","Type":"ContainerDied","Data":"886b816e510ca0a1d185b2cae870d503dc001639fe916799eb14678839db73e5"} Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.263050 4818 scope.go:117] "RemoveContainer" containerID="2011cd97d070b8a754b3acec1fbdf7b22170264b4f513d61217568bcbbd91c82" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.298616 4818 scope.go:117] "RemoveContainer" containerID="7b521d1554430e15d45e81217bcbedffcbf9f91f255aee93a639d50189365d64" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.316375 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.331757 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.344412 4818 scope.go:117] "RemoveContainer" containerID="2011cd97d070b8a754b3acec1fbdf7b22170264b4f513d61217568bcbbd91c82" Nov 22 05:14:41 crc kubenswrapper[4818]: E1122 05:14:41.344814 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2011cd97d070b8a754b3acec1fbdf7b22170264b4f513d61217568bcbbd91c82\": container with ID starting with 2011cd97d070b8a754b3acec1fbdf7b22170264b4f513d61217568bcbbd91c82 not found: ID does not exist" containerID="2011cd97d070b8a754b3acec1fbdf7b22170264b4f513d61217568bcbbd91c82" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.344856 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2011cd97d070b8a754b3acec1fbdf7b22170264b4f513d61217568bcbbd91c82"} err="failed to get container status \"2011cd97d070b8a754b3acec1fbdf7b22170264b4f513d61217568bcbbd91c82\": rpc error: code = NotFound desc = could not find container \"2011cd97d070b8a754b3acec1fbdf7b22170264b4f513d61217568bcbbd91c82\": container with ID starting with 2011cd97d070b8a754b3acec1fbdf7b22170264b4f513d61217568bcbbd91c82 not found: ID does not exist" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.344882 4818 scope.go:117] "RemoveContainer" containerID="7b521d1554430e15d45e81217bcbedffcbf9f91f255aee93a639d50189365d64" Nov 22 05:14:41 crc kubenswrapper[4818]: E1122 05:14:41.346507 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b521d1554430e15d45e81217bcbedffcbf9f91f255aee93a639d50189365d64\": container with ID starting with 7b521d1554430e15d45e81217bcbedffcbf9f91f255aee93a639d50189365d64 not found: ID does not exist" containerID="7b521d1554430e15d45e81217bcbedffcbf9f91f255aee93a639d50189365d64" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.346541 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b521d1554430e15d45e81217bcbedffcbf9f91f255aee93a639d50189365d64"} err="failed to get container status \"7b521d1554430e15d45e81217bcbedffcbf9f91f255aee93a639d50189365d64\": rpc error: code = NotFound desc = could not find container \"7b521d1554430e15d45e81217bcbedffcbf9f91f255aee93a639d50189365d64\": container with ID starting with 7b521d1554430e15d45e81217bcbedffcbf9f91f255aee93a639d50189365d64 not found: ID does not exist" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.350289 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 05:14:41 crc kubenswrapper[4818]: E1122 05:14:41.350704 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67e09220-fd61-4785-9de7-90b3ffaf0157" containerName="rabbitmq" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.350717 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="67e09220-fd61-4785-9de7-90b3ffaf0157" containerName="rabbitmq" Nov 22 05:14:41 crc kubenswrapper[4818]: E1122 05:14:41.350731 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67e09220-fd61-4785-9de7-90b3ffaf0157" containerName="setup-container" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.350737 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="67e09220-fd61-4785-9de7-90b3ffaf0157" containerName="setup-container" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.350907 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="67e09220-fd61-4785-9de7-90b3ffaf0157" containerName="rabbitmq" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.359050 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.361658 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.361865 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.361956 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.362104 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-cws5h" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.362112 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.362858 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.362883 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.363763 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.444966 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.445328 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt29k\" (UniqueName: \"kubernetes.io/projected/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-kube-api-access-bt29k\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.445356 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.445389 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.445418 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.445452 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.445491 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.445514 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.445535 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.445586 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.445661 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.586798 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.586844 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.586875 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.586892 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.586908 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.586944 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.586997 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.587024 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.587070 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt29k\" (UniqueName: \"kubernetes.io/projected/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-kube-api-access-bt29k\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.587086 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.587107 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.587319 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.589595 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.590328 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.590360 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.590517 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.591060 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.596601 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.597117 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.597117 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.601434 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.614244 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt29k\" (UniqueName: \"kubernetes.io/projected/3e1d70d4-6ada-47bf-a7f0-086ac6b891a0-kube-api-access-bt29k\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.623864 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.638736 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 05:14:41 crc kubenswrapper[4818]: I1122 05:14:41.681430 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:14:42 crc kubenswrapper[4818]: I1122 05:14:42.127045 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 05:14:42 crc kubenswrapper[4818]: W1122 05:14:42.127501 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e1d70d4_6ada_47bf_a7f0_086ac6b891a0.slice/crio-2fcf76a97c577d987fa19c20941bed4a67b3a86c7b84292a665cc830efcba86b WatchSource:0}: Error finding container 2fcf76a97c577d987fa19c20941bed4a67b3a86c7b84292a665cc830efcba86b: Status 404 returned error can't find the container with id 2fcf76a97c577d987fa19c20941bed4a67b3a86c7b84292a665cc830efcba86b Nov 22 05:14:42 crc kubenswrapper[4818]: I1122 05:14:42.274698 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0","Type":"ContainerStarted","Data":"2fcf76a97c577d987fa19c20941bed4a67b3a86c7b84292a665cc830efcba86b"} Nov 22 05:14:42 crc kubenswrapper[4818]: I1122 05:14:42.278738 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86","Type":"ContainerStarted","Data":"0c93af7e1b10ef98a529e60649e9605e395eb6a91c73ff3ea4d7e5b3859082c6"} Nov 22 05:14:42 crc kubenswrapper[4818]: I1122 05:14:42.326722 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67e09220-fd61-4785-9de7-90b3ffaf0157" path="/var/lib/kubelet/pods/67e09220-fd61-4785-9de7-90b3ffaf0157/volumes" Nov 22 05:14:42 crc kubenswrapper[4818]: I1122 05:14:42.328477 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69f72526-235a-4079-a089-43e1b93aef9f" path="/var/lib/kubelet/pods/69f72526-235a-4079-a089-43e1b93aef9f/volumes" Nov 22 05:14:44 crc kubenswrapper[4818]: I1122 05:14:44.313799 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86","Type":"ContainerStarted","Data":"4fb8aa0df33d0be753433b6d8ef39b67064fedd432064bd85f2a933475f37e66"} Nov 22 05:14:44 crc kubenswrapper[4818]: I1122 05:14:44.314320 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0","Type":"ContainerStarted","Data":"3efb886eb71e14d234ae433d2614ab3891ff4320033056bcd5e06f57273a31b3"} Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.197049 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-zvd4c"] Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.198961 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.202138 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.215902 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-zvd4c"] Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.358925 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-config\") pod \"dnsmasq-dns-6447ccbd8f-zvd4c\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.359292 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-ovsdbserver-nb\") pod \"dnsmasq-dns-6447ccbd8f-zvd4c\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.359361 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-dns-svc\") pod \"dnsmasq-dns-6447ccbd8f-zvd4c\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.359377 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-openstack-edpm-ipam\") pod \"dnsmasq-dns-6447ccbd8f-zvd4c\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.359438 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dd62n\" (UniqueName: \"kubernetes.io/projected/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-kube-api-access-dd62n\") pod \"dnsmasq-dns-6447ccbd8f-zvd4c\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.359481 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-ovsdbserver-sb\") pod \"dnsmasq-dns-6447ccbd8f-zvd4c\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.461591 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dd62n\" (UniqueName: \"kubernetes.io/projected/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-kube-api-access-dd62n\") pod \"dnsmasq-dns-6447ccbd8f-zvd4c\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.461673 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-ovsdbserver-sb\") pod \"dnsmasq-dns-6447ccbd8f-zvd4c\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.462448 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-config\") pod \"dnsmasq-dns-6447ccbd8f-zvd4c\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.462817 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-ovsdbserver-sb\") pod \"dnsmasq-dns-6447ccbd8f-zvd4c\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.463317 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-config\") pod \"dnsmasq-dns-6447ccbd8f-zvd4c\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.463531 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-ovsdbserver-nb\") pod \"dnsmasq-dns-6447ccbd8f-zvd4c\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.463630 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-dns-svc\") pod \"dnsmasq-dns-6447ccbd8f-zvd4c\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.463659 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-openstack-edpm-ipam\") pod \"dnsmasq-dns-6447ccbd8f-zvd4c\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.465088 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-ovsdbserver-nb\") pod \"dnsmasq-dns-6447ccbd8f-zvd4c\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.466480 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-dns-svc\") pod \"dnsmasq-dns-6447ccbd8f-zvd4c\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.466497 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-openstack-edpm-ipam\") pod \"dnsmasq-dns-6447ccbd8f-zvd4c\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.485177 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dd62n\" (UniqueName: \"kubernetes.io/projected/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-kube-api-access-dd62n\") pod \"dnsmasq-dns-6447ccbd8f-zvd4c\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.531422 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:45 crc kubenswrapper[4818]: I1122 05:14:45.992523 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-zvd4c"] Nov 22 05:14:45 crc kubenswrapper[4818]: W1122 05:14:45.995543 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8b2dcfc_1ca6_4ee3_bc79_26e74944d76f.slice/crio-20307077b90b4613d276879a0003cd6e1a8e82cac6c42cf6d4b7ed6d38f38a1f WatchSource:0}: Error finding container 20307077b90b4613d276879a0003cd6e1a8e82cac6c42cf6d4b7ed6d38f38a1f: Status 404 returned error can't find the container with id 20307077b90b4613d276879a0003cd6e1a8e82cac6c42cf6d4b7ed6d38f38a1f Nov 22 05:14:46 crc kubenswrapper[4818]: I1122 05:14:46.340559 4818 generic.go:334] "Generic (PLEG): container finished" podID="d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f" containerID="7c227959c8eedfbb9e337393ec88b1c6f2cf7e6203debba75c746dae1004b069" exitCode=0 Nov 22 05:14:46 crc kubenswrapper[4818]: I1122 05:14:46.340648 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" event={"ID":"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f","Type":"ContainerDied","Data":"7c227959c8eedfbb9e337393ec88b1c6f2cf7e6203debba75c746dae1004b069"} Nov 22 05:14:46 crc kubenswrapper[4818]: I1122 05:14:46.340865 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" event={"ID":"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f","Type":"ContainerStarted","Data":"20307077b90b4613d276879a0003cd6e1a8e82cac6c42cf6d4b7ed6d38f38a1f"} Nov 22 05:14:47 crc kubenswrapper[4818]: I1122 05:14:47.354810 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" event={"ID":"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f","Type":"ContainerStarted","Data":"09d3a921e87dbfd0c0ac1db1fa0d57bdee9ada26dbcd3ab243751b335201e1e1"} Nov 22 05:14:47 crc kubenswrapper[4818]: I1122 05:14:47.355095 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:47 crc kubenswrapper[4818]: I1122 05:14:47.386379 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" podStartSLOduration=2.386126703 podStartE2EDuration="2.386126703s" podCreationTimestamp="2025-11-22 05:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:14:47.37548423 +0000 UTC m=+1639.949900837" watchObservedRunningTime="2025-11-22 05:14:47.386126703 +0000 UTC m=+1639.960543260" Nov 22 05:14:51 crc kubenswrapper[4818]: I1122 05:14:51.264373 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:14:51 crc kubenswrapper[4818]: I1122 05:14:51.264800 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:14:51 crc kubenswrapper[4818]: I1122 05:14:51.264863 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 05:14:51 crc kubenswrapper[4818]: I1122 05:14:51.265876 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522"} pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 05:14:51 crc kubenswrapper[4818]: I1122 05:14:51.265973 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" containerID="cri-o://9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" gracePeriod=600 Nov 22 05:14:51 crc kubenswrapper[4818]: E1122 05:14:51.396051 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:14:51 crc kubenswrapper[4818]: I1122 05:14:51.396760 4818 generic.go:334] "Generic (PLEG): container finished" podID="0971c9ec-29ce-44eb-9875-ad4664143305" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" exitCode=0 Nov 22 05:14:51 crc kubenswrapper[4818]: I1122 05:14:51.396822 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerDied","Data":"9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522"} Nov 22 05:14:51 crc kubenswrapper[4818]: I1122 05:14:51.396879 4818 scope.go:117] "RemoveContainer" containerID="f815386123126fabb0fbbe54df8d571044020c0a205a877246cd025618c83140" Nov 22 05:14:52 crc kubenswrapper[4818]: I1122 05:14:52.408900 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:14:52 crc kubenswrapper[4818]: E1122 05:14:52.409670 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.533524 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.601661 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-88nx2"] Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.602476 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b856c5697-88nx2" podUID="4ca5e50d-5880-40a9-a08b-d8c22ad64e0b" containerName="dnsmasq-dns" containerID="cri-o://6de19732be39a975d665120b9409240a0ac7f79dee458e2e995c4b2ae325abc5" gracePeriod=10 Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.747371 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-6xc6s"] Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.748956 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.756902 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-6xc6s"] Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.864825 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-dns-svc\") pod \"dnsmasq-dns-864d5fc68c-6xc6s\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.864875 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-openstack-edpm-ipam\") pod \"dnsmasq-dns-864d5fc68c-6xc6s\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.864917 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-ovsdbserver-sb\") pod \"dnsmasq-dns-864d5fc68c-6xc6s\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.865013 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-config\") pod \"dnsmasq-dns-864d5fc68c-6xc6s\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.865035 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bchlt\" (UniqueName: \"kubernetes.io/projected/ac91350e-274b-422f-aada-5c4f42800ffd-kube-api-access-bchlt\") pod \"dnsmasq-dns-864d5fc68c-6xc6s\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.865203 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-ovsdbserver-nb\") pod \"dnsmasq-dns-864d5fc68c-6xc6s\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.967141 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-config\") pod \"dnsmasq-dns-864d5fc68c-6xc6s\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.967199 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bchlt\" (UniqueName: \"kubernetes.io/projected/ac91350e-274b-422f-aada-5c4f42800ffd-kube-api-access-bchlt\") pod \"dnsmasq-dns-864d5fc68c-6xc6s\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.967243 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-ovsdbserver-nb\") pod \"dnsmasq-dns-864d5fc68c-6xc6s\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.967324 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-dns-svc\") pod \"dnsmasq-dns-864d5fc68c-6xc6s\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.967351 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-openstack-edpm-ipam\") pod \"dnsmasq-dns-864d5fc68c-6xc6s\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.967392 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-ovsdbserver-sb\") pod \"dnsmasq-dns-864d5fc68c-6xc6s\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.968809 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-openstack-edpm-ipam\") pod \"dnsmasq-dns-864d5fc68c-6xc6s\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.968841 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-dns-svc\") pod \"dnsmasq-dns-864d5fc68c-6xc6s\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.968947 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-config\") pod \"dnsmasq-dns-864d5fc68c-6xc6s\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.969299 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-ovsdbserver-nb\") pod \"dnsmasq-dns-864d5fc68c-6xc6s\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:55 crc kubenswrapper[4818]: I1122 05:14:55.969725 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-ovsdbserver-sb\") pod \"dnsmasq-dns-864d5fc68c-6xc6s\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.000285 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bchlt\" (UniqueName: \"kubernetes.io/projected/ac91350e-274b-422f-aada-5c4f42800ffd-kube-api-access-bchlt\") pod \"dnsmasq-dns-864d5fc68c-6xc6s\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.076047 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.084275 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.169963 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-ovsdbserver-sb\") pod \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\" (UID: \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\") " Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.170011 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-config\") pod \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\" (UID: \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\") " Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.170124 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qz776\" (UniqueName: \"kubernetes.io/projected/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-kube-api-access-qz776\") pod \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\" (UID: \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\") " Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.170169 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-ovsdbserver-nb\") pod \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\" (UID: \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\") " Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.170213 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-dns-svc\") pod \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\" (UID: \"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b\") " Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.175884 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-kube-api-access-qz776" (OuterVolumeSpecName: "kube-api-access-qz776") pod "4ca5e50d-5880-40a9-a08b-d8c22ad64e0b" (UID: "4ca5e50d-5880-40a9-a08b-d8c22ad64e0b"). InnerVolumeSpecName "kube-api-access-qz776". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.254674 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4ca5e50d-5880-40a9-a08b-d8c22ad64e0b" (UID: "4ca5e50d-5880-40a9-a08b-d8c22ad64e0b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.254939 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-config" (OuterVolumeSpecName: "config") pod "4ca5e50d-5880-40a9-a08b-d8c22ad64e0b" (UID: "4ca5e50d-5880-40a9-a08b-d8c22ad64e0b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.257283 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4ca5e50d-5880-40a9-a08b-d8c22ad64e0b" (UID: "4ca5e50d-5880-40a9-a08b-d8c22ad64e0b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.270116 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4ca5e50d-5880-40a9-a08b-d8c22ad64e0b" (UID: "4ca5e50d-5880-40a9-a08b-d8c22ad64e0b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.272227 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qz776\" (UniqueName: \"kubernetes.io/projected/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-kube-api-access-qz776\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.272381 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.272396 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.272406 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.272415 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.445011 4818 generic.go:334] "Generic (PLEG): container finished" podID="4ca5e50d-5880-40a9-a08b-d8c22ad64e0b" containerID="6de19732be39a975d665120b9409240a0ac7f79dee458e2e995c4b2ae325abc5" exitCode=0 Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.445063 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-88nx2" event={"ID":"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b","Type":"ContainerDied","Data":"6de19732be39a975d665120b9409240a0ac7f79dee458e2e995c4b2ae325abc5"} Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.445073 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-88nx2" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.445096 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-88nx2" event={"ID":"4ca5e50d-5880-40a9-a08b-d8c22ad64e0b","Type":"ContainerDied","Data":"702fcd598776c9708585339c988b081837b060ec6085bd9ff4b8919251ceaca0"} Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.445120 4818 scope.go:117] "RemoveContainer" containerID="6de19732be39a975d665120b9409240a0ac7f79dee458e2e995c4b2ae325abc5" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.470749 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-88nx2"] Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.475043 4818 scope.go:117] "RemoveContainer" containerID="65a098c1be181f5c1aa80954953218c68c73bcd173598839fc40c230b11e130d" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.478530 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-88nx2"] Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.496003 4818 scope.go:117] "RemoveContainer" containerID="6de19732be39a975d665120b9409240a0ac7f79dee458e2e995c4b2ae325abc5" Nov 22 05:14:56 crc kubenswrapper[4818]: E1122 05:14:56.496573 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6de19732be39a975d665120b9409240a0ac7f79dee458e2e995c4b2ae325abc5\": container with ID starting with 6de19732be39a975d665120b9409240a0ac7f79dee458e2e995c4b2ae325abc5 not found: ID does not exist" containerID="6de19732be39a975d665120b9409240a0ac7f79dee458e2e995c4b2ae325abc5" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.496611 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6de19732be39a975d665120b9409240a0ac7f79dee458e2e995c4b2ae325abc5"} err="failed to get container status \"6de19732be39a975d665120b9409240a0ac7f79dee458e2e995c4b2ae325abc5\": rpc error: code = NotFound desc = could not find container \"6de19732be39a975d665120b9409240a0ac7f79dee458e2e995c4b2ae325abc5\": container with ID starting with 6de19732be39a975d665120b9409240a0ac7f79dee458e2e995c4b2ae325abc5 not found: ID does not exist" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.496652 4818 scope.go:117] "RemoveContainer" containerID="65a098c1be181f5c1aa80954953218c68c73bcd173598839fc40c230b11e130d" Nov 22 05:14:56 crc kubenswrapper[4818]: E1122 05:14:56.497327 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65a098c1be181f5c1aa80954953218c68c73bcd173598839fc40c230b11e130d\": container with ID starting with 65a098c1be181f5c1aa80954953218c68c73bcd173598839fc40c230b11e130d not found: ID does not exist" containerID="65a098c1be181f5c1aa80954953218c68c73bcd173598839fc40c230b11e130d" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.497365 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65a098c1be181f5c1aa80954953218c68c73bcd173598839fc40c230b11e130d"} err="failed to get container status \"65a098c1be181f5c1aa80954953218c68c73bcd173598839fc40c230b11e130d\": rpc error: code = NotFound desc = could not find container \"65a098c1be181f5c1aa80954953218c68c73bcd173598839fc40c230b11e130d\": container with ID starting with 65a098c1be181f5c1aa80954953218c68c73bcd173598839fc40c230b11e130d not found: ID does not exist" Nov 22 05:14:56 crc kubenswrapper[4818]: I1122 05:14:56.603744 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-6xc6s"] Nov 22 05:14:56 crc kubenswrapper[4818]: W1122 05:14:56.612586 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac91350e_274b_422f_aada_5c4f42800ffd.slice/crio-1d90dfd004e65507b32c100c598435d8882baffb8e04b296691b45f061eeccb4 WatchSource:0}: Error finding container 1d90dfd004e65507b32c100c598435d8882baffb8e04b296691b45f061eeccb4: Status 404 returned error can't find the container with id 1d90dfd004e65507b32c100c598435d8882baffb8e04b296691b45f061eeccb4 Nov 22 05:14:57 crc kubenswrapper[4818]: I1122 05:14:57.460505 4818 generic.go:334] "Generic (PLEG): container finished" podID="ac91350e-274b-422f-aada-5c4f42800ffd" containerID="f9e5ade1c43acf64c258edc95ee2f2bc24022eceb0dadc49922d0961b9087223" exitCode=0 Nov 22 05:14:57 crc kubenswrapper[4818]: I1122 05:14:57.460614 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" event={"ID":"ac91350e-274b-422f-aada-5c4f42800ffd","Type":"ContainerDied","Data":"f9e5ade1c43acf64c258edc95ee2f2bc24022eceb0dadc49922d0961b9087223"} Nov 22 05:14:57 crc kubenswrapper[4818]: I1122 05:14:57.460877 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" event={"ID":"ac91350e-274b-422f-aada-5c4f42800ffd","Type":"ContainerStarted","Data":"1d90dfd004e65507b32c100c598435d8882baffb8e04b296691b45f061eeccb4"} Nov 22 05:14:58 crc kubenswrapper[4818]: I1122 05:14:58.328240 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ca5e50d-5880-40a9-a08b-d8c22ad64e0b" path="/var/lib/kubelet/pods/4ca5e50d-5880-40a9-a08b-d8c22ad64e0b/volumes" Nov 22 05:14:58 crc kubenswrapper[4818]: I1122 05:14:58.470819 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" event={"ID":"ac91350e-274b-422f-aada-5c4f42800ffd","Type":"ContainerStarted","Data":"4a6bd71c9a869a93ca44b3b32c421e5cf41ce27e9689ad6cdf9a4e39b64dabfb"} Nov 22 05:14:58 crc kubenswrapper[4818]: I1122 05:14:58.470999 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:14:58 crc kubenswrapper[4818]: I1122 05:14:58.489275 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" podStartSLOduration=3.489242075 podStartE2EDuration="3.489242075s" podCreationTimestamp="2025-11-22 05:14:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:14:58.488192437 +0000 UTC m=+1651.062608974" watchObservedRunningTime="2025-11-22 05:14:58.489242075 +0000 UTC m=+1651.063658612" Nov 22 05:14:59 crc kubenswrapper[4818]: I1122 05:14:59.545904 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fd6vn"] Nov 22 05:14:59 crc kubenswrapper[4818]: E1122 05:14:59.546883 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ca5e50d-5880-40a9-a08b-d8c22ad64e0b" containerName="init" Nov 22 05:14:59 crc kubenswrapper[4818]: I1122 05:14:59.546906 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ca5e50d-5880-40a9-a08b-d8c22ad64e0b" containerName="init" Nov 22 05:14:59 crc kubenswrapper[4818]: E1122 05:14:59.546922 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ca5e50d-5880-40a9-a08b-d8c22ad64e0b" containerName="dnsmasq-dns" Nov 22 05:14:59 crc kubenswrapper[4818]: I1122 05:14:59.546933 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ca5e50d-5880-40a9-a08b-d8c22ad64e0b" containerName="dnsmasq-dns" Nov 22 05:14:59 crc kubenswrapper[4818]: I1122 05:14:59.547335 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ca5e50d-5880-40a9-a08b-d8c22ad64e0b" containerName="dnsmasq-dns" Nov 22 05:14:59 crc kubenswrapper[4818]: I1122 05:14:59.549660 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fd6vn" Nov 22 05:14:59 crc kubenswrapper[4818]: I1122 05:14:59.555315 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fd6vn"] Nov 22 05:14:59 crc kubenswrapper[4818]: I1122 05:14:59.641819 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f4c1446-a45c-4230-904b-807d441161e7-catalog-content\") pod \"redhat-marketplace-fd6vn\" (UID: \"7f4c1446-a45c-4230-904b-807d441161e7\") " pod="openshift-marketplace/redhat-marketplace-fd6vn" Nov 22 05:14:59 crc kubenswrapper[4818]: I1122 05:14:59.641901 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f4c1446-a45c-4230-904b-807d441161e7-utilities\") pod \"redhat-marketplace-fd6vn\" (UID: \"7f4c1446-a45c-4230-904b-807d441161e7\") " pod="openshift-marketplace/redhat-marketplace-fd6vn" Nov 22 05:14:59 crc kubenswrapper[4818]: I1122 05:14:59.641999 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8r8b\" (UniqueName: \"kubernetes.io/projected/7f4c1446-a45c-4230-904b-807d441161e7-kube-api-access-t8r8b\") pod \"redhat-marketplace-fd6vn\" (UID: \"7f4c1446-a45c-4230-904b-807d441161e7\") " pod="openshift-marketplace/redhat-marketplace-fd6vn" Nov 22 05:14:59 crc kubenswrapper[4818]: I1122 05:14:59.743816 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f4c1446-a45c-4230-904b-807d441161e7-utilities\") pod \"redhat-marketplace-fd6vn\" (UID: \"7f4c1446-a45c-4230-904b-807d441161e7\") " pod="openshift-marketplace/redhat-marketplace-fd6vn" Nov 22 05:14:59 crc kubenswrapper[4818]: I1122 05:14:59.743971 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8r8b\" (UniqueName: \"kubernetes.io/projected/7f4c1446-a45c-4230-904b-807d441161e7-kube-api-access-t8r8b\") pod \"redhat-marketplace-fd6vn\" (UID: \"7f4c1446-a45c-4230-904b-807d441161e7\") " pod="openshift-marketplace/redhat-marketplace-fd6vn" Nov 22 05:14:59 crc kubenswrapper[4818]: I1122 05:14:59.744208 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f4c1446-a45c-4230-904b-807d441161e7-catalog-content\") pod \"redhat-marketplace-fd6vn\" (UID: \"7f4c1446-a45c-4230-904b-807d441161e7\") " pod="openshift-marketplace/redhat-marketplace-fd6vn" Nov 22 05:14:59 crc kubenswrapper[4818]: I1122 05:14:59.744406 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f4c1446-a45c-4230-904b-807d441161e7-utilities\") pod \"redhat-marketplace-fd6vn\" (UID: \"7f4c1446-a45c-4230-904b-807d441161e7\") " pod="openshift-marketplace/redhat-marketplace-fd6vn" Nov 22 05:14:59 crc kubenswrapper[4818]: I1122 05:14:59.744911 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f4c1446-a45c-4230-904b-807d441161e7-catalog-content\") pod \"redhat-marketplace-fd6vn\" (UID: \"7f4c1446-a45c-4230-904b-807d441161e7\") " pod="openshift-marketplace/redhat-marketplace-fd6vn" Nov 22 05:14:59 crc kubenswrapper[4818]: I1122 05:14:59.763879 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8r8b\" (UniqueName: \"kubernetes.io/projected/7f4c1446-a45c-4230-904b-807d441161e7-kube-api-access-t8r8b\") pod \"redhat-marketplace-fd6vn\" (UID: \"7f4c1446-a45c-4230-904b-807d441161e7\") " pod="openshift-marketplace/redhat-marketplace-fd6vn" Nov 22 05:14:59 crc kubenswrapper[4818]: I1122 05:14:59.873116 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fd6vn" Nov 22 05:15:00 crc kubenswrapper[4818]: I1122 05:15:00.156370 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w"] Nov 22 05:15:00 crc kubenswrapper[4818]: I1122 05:15:00.158047 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w" Nov 22 05:15:00 crc kubenswrapper[4818]: I1122 05:15:00.162043 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 05:15:00 crc kubenswrapper[4818]: I1122 05:15:00.165492 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 05:15:00 crc kubenswrapper[4818]: I1122 05:15:00.175333 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w"] Nov 22 05:15:00 crc kubenswrapper[4818]: I1122 05:15:00.251529 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/33d93220-7194-4413-9d83-3eb08fe261d4-config-volume\") pod \"collect-profiles-29396475-zcs8w\" (UID: \"33d93220-7194-4413-9d83-3eb08fe261d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w" Nov 22 05:15:00 crc kubenswrapper[4818]: I1122 05:15:00.252036 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phtvc\" (UniqueName: \"kubernetes.io/projected/33d93220-7194-4413-9d83-3eb08fe261d4-kube-api-access-phtvc\") pod \"collect-profiles-29396475-zcs8w\" (UID: \"33d93220-7194-4413-9d83-3eb08fe261d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w" Nov 22 05:15:00 crc kubenswrapper[4818]: I1122 05:15:00.252275 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/33d93220-7194-4413-9d83-3eb08fe261d4-secret-volume\") pod \"collect-profiles-29396475-zcs8w\" (UID: \"33d93220-7194-4413-9d83-3eb08fe261d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w" Nov 22 05:15:00 crc kubenswrapper[4818]: I1122 05:15:00.352590 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fd6vn"] Nov 22 05:15:00 crc kubenswrapper[4818]: I1122 05:15:00.354569 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/33d93220-7194-4413-9d83-3eb08fe261d4-secret-volume\") pod \"collect-profiles-29396475-zcs8w\" (UID: \"33d93220-7194-4413-9d83-3eb08fe261d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w" Nov 22 05:15:00 crc kubenswrapper[4818]: I1122 05:15:00.354637 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/33d93220-7194-4413-9d83-3eb08fe261d4-config-volume\") pod \"collect-profiles-29396475-zcs8w\" (UID: \"33d93220-7194-4413-9d83-3eb08fe261d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w" Nov 22 05:15:00 crc kubenswrapper[4818]: I1122 05:15:00.354953 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phtvc\" (UniqueName: \"kubernetes.io/projected/33d93220-7194-4413-9d83-3eb08fe261d4-kube-api-access-phtvc\") pod \"collect-profiles-29396475-zcs8w\" (UID: \"33d93220-7194-4413-9d83-3eb08fe261d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w" Nov 22 05:15:00 crc kubenswrapper[4818]: I1122 05:15:00.365270 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/33d93220-7194-4413-9d83-3eb08fe261d4-secret-volume\") pod \"collect-profiles-29396475-zcs8w\" (UID: \"33d93220-7194-4413-9d83-3eb08fe261d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w" Nov 22 05:15:00 crc kubenswrapper[4818]: I1122 05:15:00.367478 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/33d93220-7194-4413-9d83-3eb08fe261d4-config-volume\") pod \"collect-profiles-29396475-zcs8w\" (UID: \"33d93220-7194-4413-9d83-3eb08fe261d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w" Nov 22 05:15:00 crc kubenswrapper[4818]: I1122 05:15:00.375423 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phtvc\" (UniqueName: \"kubernetes.io/projected/33d93220-7194-4413-9d83-3eb08fe261d4-kube-api-access-phtvc\") pod \"collect-profiles-29396475-zcs8w\" (UID: \"33d93220-7194-4413-9d83-3eb08fe261d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w" Nov 22 05:15:00 crc kubenswrapper[4818]: I1122 05:15:00.480491 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w" Nov 22 05:15:00 crc kubenswrapper[4818]: I1122 05:15:00.498177 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fd6vn" event={"ID":"7f4c1446-a45c-4230-904b-807d441161e7","Type":"ContainerStarted","Data":"c141f6edf0ec69a38dab02ae83ebd169364c6172881bfb29e3608ef77cedd5da"} Nov 22 05:15:00 crc kubenswrapper[4818]: I1122 05:15:00.905813 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w"] Nov 22 05:15:00 crc kubenswrapper[4818]: W1122 05:15:00.908861 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33d93220_7194_4413_9d83_3eb08fe261d4.slice/crio-aef62a54dabd2457c16f30379fe5311ea4a01b584979d08cda28d216195bf940 WatchSource:0}: Error finding container aef62a54dabd2457c16f30379fe5311ea4a01b584979d08cda28d216195bf940: Status 404 returned error can't find the container with id aef62a54dabd2457c16f30379fe5311ea4a01b584979d08cda28d216195bf940 Nov 22 05:15:00 crc kubenswrapper[4818]: I1122 05:15:00.932273 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b856c5697-88nx2" podUID="4ca5e50d-5880-40a9-a08b-d8c22ad64e0b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.180:5353: i/o timeout" Nov 22 05:15:01 crc kubenswrapper[4818]: I1122 05:15:01.509381 4818 generic.go:334] "Generic (PLEG): container finished" podID="33d93220-7194-4413-9d83-3eb08fe261d4" containerID="33b0c0a8d76db88f05bf6e94e1d03cce8a2ce5ec1aaf25dd070ac8633ed03fd2" exitCode=0 Nov 22 05:15:01 crc kubenswrapper[4818]: I1122 05:15:01.509484 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w" event={"ID":"33d93220-7194-4413-9d83-3eb08fe261d4","Type":"ContainerDied","Data":"33b0c0a8d76db88f05bf6e94e1d03cce8a2ce5ec1aaf25dd070ac8633ed03fd2"} Nov 22 05:15:01 crc kubenswrapper[4818]: I1122 05:15:01.509704 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w" event={"ID":"33d93220-7194-4413-9d83-3eb08fe261d4","Type":"ContainerStarted","Data":"aef62a54dabd2457c16f30379fe5311ea4a01b584979d08cda28d216195bf940"} Nov 22 05:15:01 crc kubenswrapper[4818]: I1122 05:15:01.511398 4818 generic.go:334] "Generic (PLEG): container finished" podID="7f4c1446-a45c-4230-904b-807d441161e7" containerID="3e29f82a3608d070beacbd177bb9c2c337e28393926622e6aff6028f99b4f3af" exitCode=0 Nov 22 05:15:01 crc kubenswrapper[4818]: I1122 05:15:01.511428 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fd6vn" event={"ID":"7f4c1446-a45c-4230-904b-807d441161e7","Type":"ContainerDied","Data":"3e29f82a3608d070beacbd177bb9c2c337e28393926622e6aff6028f99b4f3af"} Nov 22 05:15:02 crc kubenswrapper[4818]: I1122 05:15:02.527218 4818 generic.go:334] "Generic (PLEG): container finished" podID="7f4c1446-a45c-4230-904b-807d441161e7" containerID="cc82de1e1f3b9ad0a9d07c7898650887087183395cef5716e9816528e699926c" exitCode=0 Nov 22 05:15:02 crc kubenswrapper[4818]: I1122 05:15:02.527318 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fd6vn" event={"ID":"7f4c1446-a45c-4230-904b-807d441161e7","Type":"ContainerDied","Data":"cc82de1e1f3b9ad0a9d07c7898650887087183395cef5716e9816528e699926c"} Nov 22 05:15:02 crc kubenswrapper[4818]: I1122 05:15:02.908512 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w" Nov 22 05:15:03 crc kubenswrapper[4818]: I1122 05:15:03.011505 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phtvc\" (UniqueName: \"kubernetes.io/projected/33d93220-7194-4413-9d83-3eb08fe261d4-kube-api-access-phtvc\") pod \"33d93220-7194-4413-9d83-3eb08fe261d4\" (UID: \"33d93220-7194-4413-9d83-3eb08fe261d4\") " Nov 22 05:15:03 crc kubenswrapper[4818]: I1122 05:15:03.011586 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/33d93220-7194-4413-9d83-3eb08fe261d4-config-volume\") pod \"33d93220-7194-4413-9d83-3eb08fe261d4\" (UID: \"33d93220-7194-4413-9d83-3eb08fe261d4\") " Nov 22 05:15:03 crc kubenswrapper[4818]: I1122 05:15:03.011625 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/33d93220-7194-4413-9d83-3eb08fe261d4-secret-volume\") pod \"33d93220-7194-4413-9d83-3eb08fe261d4\" (UID: \"33d93220-7194-4413-9d83-3eb08fe261d4\") " Nov 22 05:15:03 crc kubenswrapper[4818]: I1122 05:15:03.013777 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33d93220-7194-4413-9d83-3eb08fe261d4-config-volume" (OuterVolumeSpecName: "config-volume") pod "33d93220-7194-4413-9d83-3eb08fe261d4" (UID: "33d93220-7194-4413-9d83-3eb08fe261d4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:15:03 crc kubenswrapper[4818]: I1122 05:15:03.018666 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33d93220-7194-4413-9d83-3eb08fe261d4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "33d93220-7194-4413-9d83-3eb08fe261d4" (UID: "33d93220-7194-4413-9d83-3eb08fe261d4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:15:03 crc kubenswrapper[4818]: I1122 05:15:03.018717 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33d93220-7194-4413-9d83-3eb08fe261d4-kube-api-access-phtvc" (OuterVolumeSpecName: "kube-api-access-phtvc") pod "33d93220-7194-4413-9d83-3eb08fe261d4" (UID: "33d93220-7194-4413-9d83-3eb08fe261d4"). InnerVolumeSpecName "kube-api-access-phtvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:15:03 crc kubenswrapper[4818]: I1122 05:15:03.113710 4818 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/33d93220-7194-4413-9d83-3eb08fe261d4-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:03 crc kubenswrapper[4818]: I1122 05:15:03.113763 4818 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/33d93220-7194-4413-9d83-3eb08fe261d4-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:03 crc kubenswrapper[4818]: I1122 05:15:03.113775 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phtvc\" (UniqueName: \"kubernetes.io/projected/33d93220-7194-4413-9d83-3eb08fe261d4-kube-api-access-phtvc\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:03 crc kubenswrapper[4818]: I1122 05:15:03.538303 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w" event={"ID":"33d93220-7194-4413-9d83-3eb08fe261d4","Type":"ContainerDied","Data":"aef62a54dabd2457c16f30379fe5311ea4a01b584979d08cda28d216195bf940"} Nov 22 05:15:03 crc kubenswrapper[4818]: I1122 05:15:03.538657 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aef62a54dabd2457c16f30379fe5311ea4a01b584979d08cda28d216195bf940" Nov 22 05:15:03 crc kubenswrapper[4818]: I1122 05:15:03.538373 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w" Nov 22 05:15:03 crc kubenswrapper[4818]: I1122 05:15:03.541566 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fd6vn" event={"ID":"7f4c1446-a45c-4230-904b-807d441161e7","Type":"ContainerStarted","Data":"67c394a98d6315180ce9264355326d3a211feaa579d89657fe31ac08622bc888"} Nov 22 05:15:03 crc kubenswrapper[4818]: I1122 05:15:03.569125 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fd6vn" podStartSLOduration=3.111494246 podStartE2EDuration="4.569106525s" podCreationTimestamp="2025-11-22 05:14:59 +0000 UTC" firstStartedPulling="2025-11-22 05:15:01.512940393 +0000 UTC m=+1654.087356920" lastFinishedPulling="2025-11-22 05:15:02.970552672 +0000 UTC m=+1655.544969199" observedRunningTime="2025-11-22 05:15:03.564951724 +0000 UTC m=+1656.139368261" watchObservedRunningTime="2025-11-22 05:15:03.569106525 +0000 UTC m=+1656.143523052" Nov 22 05:15:04 crc kubenswrapper[4818]: I1122 05:15:04.291706 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:15:04 crc kubenswrapper[4818]: E1122 05:15:04.292453 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.077472 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.155822 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-zvd4c"] Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.156228 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" podUID="d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f" containerName="dnsmasq-dns" containerID="cri-o://09d3a921e87dbfd0c0ac1db1fa0d57bdee9ada26dbcd3ab243751b335201e1e1" gracePeriod=10 Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.585127 4818 generic.go:334] "Generic (PLEG): container finished" podID="d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f" containerID="09d3a921e87dbfd0c0ac1db1fa0d57bdee9ada26dbcd3ab243751b335201e1e1" exitCode=0 Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.585405 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" event={"ID":"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f","Type":"ContainerDied","Data":"09d3a921e87dbfd0c0ac1db1fa0d57bdee9ada26dbcd3ab243751b335201e1e1"} Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.585430 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" event={"ID":"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f","Type":"ContainerDied","Data":"20307077b90b4613d276879a0003cd6e1a8e82cac6c42cf6d4b7ed6d38f38a1f"} Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.585443 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20307077b90b4613d276879a0003cd6e1a8e82cac6c42cf6d4b7ed6d38f38a1f" Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.645315 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.677289 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-openstack-edpm-ipam\") pod \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.677365 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dd62n\" (UniqueName: \"kubernetes.io/projected/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-kube-api-access-dd62n\") pod \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.677416 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-ovsdbserver-sb\") pod \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.677458 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-config\") pod \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.677501 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-dns-svc\") pod \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.677551 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-ovsdbserver-nb\") pod \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\" (UID: \"d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f\") " Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.695080 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-kube-api-access-dd62n" (OuterVolumeSpecName: "kube-api-access-dd62n") pod "d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f" (UID: "d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f"). InnerVolumeSpecName "kube-api-access-dd62n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.726788 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-config" (OuterVolumeSpecName: "config") pod "d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f" (UID: "d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.729043 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f" (UID: "d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.729542 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f" (UID: "d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.733510 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f" (UID: "d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.748841 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f" (UID: "d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.780768 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.780804 4818 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.780820 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dd62n\" (UniqueName: \"kubernetes.io/projected/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-kube-api-access-dd62n\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.780834 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.780847 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:06 crc kubenswrapper[4818]: I1122 05:15:06.780860 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:07 crc kubenswrapper[4818]: I1122 05:15:07.596364 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-zvd4c" Nov 22 05:15:07 crc kubenswrapper[4818]: I1122 05:15:07.633813 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-zvd4c"] Nov 22 05:15:07 crc kubenswrapper[4818]: I1122 05:15:07.640868 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-zvd4c"] Nov 22 05:15:08 crc kubenswrapper[4818]: I1122 05:15:08.312128 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f" path="/var/lib/kubelet/pods/d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f/volumes" Nov 22 05:15:09 crc kubenswrapper[4818]: I1122 05:15:09.874220 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fd6vn" Nov 22 05:15:09 crc kubenswrapper[4818]: I1122 05:15:09.874718 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fd6vn" Nov 22 05:15:09 crc kubenswrapper[4818]: I1122 05:15:09.978022 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fd6vn" Nov 22 05:15:10 crc kubenswrapper[4818]: I1122 05:15:10.690924 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fd6vn" Nov 22 05:15:10 crc kubenswrapper[4818]: I1122 05:15:10.757461 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fd6vn"] Nov 22 05:15:12 crc kubenswrapper[4818]: I1122 05:15:12.665642 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fd6vn" podUID="7f4c1446-a45c-4230-904b-807d441161e7" containerName="registry-server" containerID="cri-o://67c394a98d6315180ce9264355326d3a211feaa579d89657fe31ac08622bc888" gracePeriod=2 Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.206444 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fd6vn" Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.311348 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f4c1446-a45c-4230-904b-807d441161e7-catalog-content\") pod \"7f4c1446-a45c-4230-904b-807d441161e7\" (UID: \"7f4c1446-a45c-4230-904b-807d441161e7\") " Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.312020 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8r8b\" (UniqueName: \"kubernetes.io/projected/7f4c1446-a45c-4230-904b-807d441161e7-kube-api-access-t8r8b\") pod \"7f4c1446-a45c-4230-904b-807d441161e7\" (UID: \"7f4c1446-a45c-4230-904b-807d441161e7\") " Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.312203 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f4c1446-a45c-4230-904b-807d441161e7-utilities\") pod \"7f4c1446-a45c-4230-904b-807d441161e7\" (UID: \"7f4c1446-a45c-4230-904b-807d441161e7\") " Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.312851 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f4c1446-a45c-4230-904b-807d441161e7-utilities" (OuterVolumeSpecName: "utilities") pod "7f4c1446-a45c-4230-904b-807d441161e7" (UID: "7f4c1446-a45c-4230-904b-807d441161e7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.313146 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f4c1446-a45c-4230-904b-807d441161e7-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.319867 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f4c1446-a45c-4230-904b-807d441161e7-kube-api-access-t8r8b" (OuterVolumeSpecName: "kube-api-access-t8r8b") pod "7f4c1446-a45c-4230-904b-807d441161e7" (UID: "7f4c1446-a45c-4230-904b-807d441161e7"). InnerVolumeSpecName "kube-api-access-t8r8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.327710 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f4c1446-a45c-4230-904b-807d441161e7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7f4c1446-a45c-4230-904b-807d441161e7" (UID: "7f4c1446-a45c-4230-904b-807d441161e7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.415315 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8r8b\" (UniqueName: \"kubernetes.io/projected/7f4c1446-a45c-4230-904b-807d441161e7-kube-api-access-t8r8b\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.415383 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f4c1446-a45c-4230-904b-807d441161e7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.680745 4818 generic.go:334] "Generic (PLEG): container finished" podID="7f4c1446-a45c-4230-904b-807d441161e7" containerID="67c394a98d6315180ce9264355326d3a211feaa579d89657fe31ac08622bc888" exitCode=0 Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.680877 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fd6vn" Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.681431 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fd6vn" event={"ID":"7f4c1446-a45c-4230-904b-807d441161e7","Type":"ContainerDied","Data":"67c394a98d6315180ce9264355326d3a211feaa579d89657fe31ac08622bc888"} Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.681499 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fd6vn" event={"ID":"7f4c1446-a45c-4230-904b-807d441161e7","Type":"ContainerDied","Data":"c141f6edf0ec69a38dab02ae83ebd169364c6172881bfb29e3608ef77cedd5da"} Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.681530 4818 scope.go:117] "RemoveContainer" containerID="67c394a98d6315180ce9264355326d3a211feaa579d89657fe31ac08622bc888" Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.727957 4818 scope.go:117] "RemoveContainer" containerID="cc82de1e1f3b9ad0a9d07c7898650887087183395cef5716e9816528e699926c" Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.740996 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fd6vn"] Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.751870 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fd6vn"] Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.770994 4818 scope.go:117] "RemoveContainer" containerID="3e29f82a3608d070beacbd177bb9c2c337e28393926622e6aff6028f99b4f3af" Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.812801 4818 scope.go:117] "RemoveContainer" containerID="67c394a98d6315180ce9264355326d3a211feaa579d89657fe31ac08622bc888" Nov 22 05:15:13 crc kubenswrapper[4818]: E1122 05:15:13.813412 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67c394a98d6315180ce9264355326d3a211feaa579d89657fe31ac08622bc888\": container with ID starting with 67c394a98d6315180ce9264355326d3a211feaa579d89657fe31ac08622bc888 not found: ID does not exist" containerID="67c394a98d6315180ce9264355326d3a211feaa579d89657fe31ac08622bc888" Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.813550 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67c394a98d6315180ce9264355326d3a211feaa579d89657fe31ac08622bc888"} err="failed to get container status \"67c394a98d6315180ce9264355326d3a211feaa579d89657fe31ac08622bc888\": rpc error: code = NotFound desc = could not find container \"67c394a98d6315180ce9264355326d3a211feaa579d89657fe31ac08622bc888\": container with ID starting with 67c394a98d6315180ce9264355326d3a211feaa579d89657fe31ac08622bc888 not found: ID does not exist" Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.813590 4818 scope.go:117] "RemoveContainer" containerID="cc82de1e1f3b9ad0a9d07c7898650887087183395cef5716e9816528e699926c" Nov 22 05:15:13 crc kubenswrapper[4818]: E1122 05:15:13.814987 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc82de1e1f3b9ad0a9d07c7898650887087183395cef5716e9816528e699926c\": container with ID starting with cc82de1e1f3b9ad0a9d07c7898650887087183395cef5716e9816528e699926c not found: ID does not exist" containerID="cc82de1e1f3b9ad0a9d07c7898650887087183395cef5716e9816528e699926c" Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.815023 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc82de1e1f3b9ad0a9d07c7898650887087183395cef5716e9816528e699926c"} err="failed to get container status \"cc82de1e1f3b9ad0a9d07c7898650887087183395cef5716e9816528e699926c\": rpc error: code = NotFound desc = could not find container \"cc82de1e1f3b9ad0a9d07c7898650887087183395cef5716e9816528e699926c\": container with ID starting with cc82de1e1f3b9ad0a9d07c7898650887087183395cef5716e9816528e699926c not found: ID does not exist" Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.815075 4818 scope.go:117] "RemoveContainer" containerID="3e29f82a3608d070beacbd177bb9c2c337e28393926622e6aff6028f99b4f3af" Nov 22 05:15:13 crc kubenswrapper[4818]: E1122 05:15:13.817678 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e29f82a3608d070beacbd177bb9c2c337e28393926622e6aff6028f99b4f3af\": container with ID starting with 3e29f82a3608d070beacbd177bb9c2c337e28393926622e6aff6028f99b4f3af not found: ID does not exist" containerID="3e29f82a3608d070beacbd177bb9c2c337e28393926622e6aff6028f99b4f3af" Nov 22 05:15:13 crc kubenswrapper[4818]: I1122 05:15:13.817725 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e29f82a3608d070beacbd177bb9c2c337e28393926622e6aff6028f99b4f3af"} err="failed to get container status \"3e29f82a3608d070beacbd177bb9c2c337e28393926622e6aff6028f99b4f3af\": rpc error: code = NotFound desc = could not find container \"3e29f82a3608d070beacbd177bb9c2c337e28393926622e6aff6028f99b4f3af\": container with ID starting with 3e29f82a3608d070beacbd177bb9c2c337e28393926622e6aff6028f99b4f3af not found: ID does not exist" Nov 22 05:15:14 crc kubenswrapper[4818]: I1122 05:15:14.321191 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f4c1446-a45c-4230-904b-807d441161e7" path="/var/lib/kubelet/pods/7f4c1446-a45c-4230-904b-807d441161e7/volumes" Nov 22 05:15:16 crc kubenswrapper[4818]: E1122 05:15:16.043472 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70fd5462_fc4d_4e67_ace7_1f4e9a2a7c86.slice/crio-4fb8aa0df33d0be753433b6d8ef39b67064fedd432064bd85f2a933475f37e66.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70fd5462_fc4d_4e67_ace7_1f4e9a2a7c86.slice/crio-conmon-4fb8aa0df33d0be753433b6d8ef39b67064fedd432064bd85f2a933475f37e66.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e1d70d4_6ada_47bf_a7f0_086ac6b891a0.slice/crio-3efb886eb71e14d234ae433d2614ab3891ff4320033056bcd5e06f57273a31b3.scope\": RecentStats: unable to find data in memory cache]" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.292115 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:15:16 crc kubenswrapper[4818]: E1122 05:15:16.292472 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.432912 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6"] Nov 22 05:15:16 crc kubenswrapper[4818]: E1122 05:15:16.433301 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f" containerName="init" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.433312 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f" containerName="init" Nov 22 05:15:16 crc kubenswrapper[4818]: E1122 05:15:16.433321 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33d93220-7194-4413-9d83-3eb08fe261d4" containerName="collect-profiles" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.433327 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="33d93220-7194-4413-9d83-3eb08fe261d4" containerName="collect-profiles" Nov 22 05:15:16 crc kubenswrapper[4818]: E1122 05:15:16.433334 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f4c1446-a45c-4230-904b-807d441161e7" containerName="extract-content" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.433339 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f4c1446-a45c-4230-904b-807d441161e7" containerName="extract-content" Nov 22 05:15:16 crc kubenswrapper[4818]: E1122 05:15:16.433356 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f" containerName="dnsmasq-dns" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.433362 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f" containerName="dnsmasq-dns" Nov 22 05:15:16 crc kubenswrapper[4818]: E1122 05:15:16.433371 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f4c1446-a45c-4230-904b-807d441161e7" containerName="extract-utilities" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.433377 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f4c1446-a45c-4230-904b-807d441161e7" containerName="extract-utilities" Nov 22 05:15:16 crc kubenswrapper[4818]: E1122 05:15:16.433386 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f4c1446-a45c-4230-904b-807d441161e7" containerName="registry-server" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.433391 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f4c1446-a45c-4230-904b-807d441161e7" containerName="registry-server" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.433577 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f4c1446-a45c-4230-904b-807d441161e7" containerName="registry-server" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.433618 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8b2dcfc-1ca6-4ee3-bc79-26e74944d76f" containerName="dnsmasq-dns" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.433633 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="33d93220-7194-4413-9d83-3eb08fe261d4" containerName="collect-profiles" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.435311 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.443434 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.443490 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.443490 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.443688 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.460968 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6"] Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.472146 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9461206a-81ad-4c30-a0fe-18ecf0502570-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6\" (UID: \"9461206a-81ad-4c30-a0fe-18ecf0502570\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.472560 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9461206a-81ad-4c30-a0fe-18ecf0502570-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6\" (UID: \"9461206a-81ad-4c30-a0fe-18ecf0502570\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.472756 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9461206a-81ad-4c30-a0fe-18ecf0502570-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6\" (UID: \"9461206a-81ad-4c30-a0fe-18ecf0502570\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.473092 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59w7v\" (UniqueName: \"kubernetes.io/projected/9461206a-81ad-4c30-a0fe-18ecf0502570-kube-api-access-59w7v\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6\" (UID: \"9461206a-81ad-4c30-a0fe-18ecf0502570\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.574414 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59w7v\" (UniqueName: \"kubernetes.io/projected/9461206a-81ad-4c30-a0fe-18ecf0502570-kube-api-access-59w7v\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6\" (UID: \"9461206a-81ad-4c30-a0fe-18ecf0502570\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.574456 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9461206a-81ad-4c30-a0fe-18ecf0502570-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6\" (UID: \"9461206a-81ad-4c30-a0fe-18ecf0502570\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.574495 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9461206a-81ad-4c30-a0fe-18ecf0502570-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6\" (UID: \"9461206a-81ad-4c30-a0fe-18ecf0502570\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.574526 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9461206a-81ad-4c30-a0fe-18ecf0502570-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6\" (UID: \"9461206a-81ad-4c30-a0fe-18ecf0502570\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.579706 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9461206a-81ad-4c30-a0fe-18ecf0502570-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6\" (UID: \"9461206a-81ad-4c30-a0fe-18ecf0502570\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.580176 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9461206a-81ad-4c30-a0fe-18ecf0502570-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6\" (UID: \"9461206a-81ad-4c30-a0fe-18ecf0502570\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.580808 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9461206a-81ad-4c30-a0fe-18ecf0502570-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6\" (UID: \"9461206a-81ad-4c30-a0fe-18ecf0502570\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.591557 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59w7v\" (UniqueName: \"kubernetes.io/projected/9461206a-81ad-4c30-a0fe-18ecf0502570-kube-api-access-59w7v\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6\" (UID: \"9461206a-81ad-4c30-a0fe-18ecf0502570\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.714300 4818 generic.go:334] "Generic (PLEG): container finished" podID="3e1d70d4-6ada-47bf-a7f0-086ac6b891a0" containerID="3efb886eb71e14d234ae433d2614ab3891ff4320033056bcd5e06f57273a31b3" exitCode=0 Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.714443 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0","Type":"ContainerDied","Data":"3efb886eb71e14d234ae433d2614ab3891ff4320033056bcd5e06f57273a31b3"} Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.716454 4818 generic.go:334] "Generic (PLEG): container finished" podID="70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86" containerID="4fb8aa0df33d0be753433b6d8ef39b67064fedd432064bd85f2a933475f37e66" exitCode=0 Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.716530 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86","Type":"ContainerDied","Data":"4fb8aa0df33d0be753433b6d8ef39b67064fedd432064bd85f2a933475f37e66"} Nov 22 05:15:16 crc kubenswrapper[4818]: I1122 05:15:16.763901 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.294645 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6"] Nov 22 05:15:17 crc kubenswrapper[4818]: W1122 05:15:17.297046 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9461206a_81ad_4c30_a0fe_18ecf0502570.slice/crio-b31a72c18b8713fc4fb1693bacb27b2bab5936ec56038c9d249f4c3fc878f2d9 WatchSource:0}: Error finding container b31a72c18b8713fc4fb1693bacb27b2bab5936ec56038c9d249f4c3fc878f2d9: Status 404 returned error can't find the container with id b31a72c18b8713fc4fb1693bacb27b2bab5936ec56038c9d249f4c3fc878f2d9 Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.646598 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jsvv8"] Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.649150 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jsvv8" Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.657477 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jsvv8"] Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.700104 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c94b4ce9-b4a7-48d6-940c-198c0e29f213-catalog-content\") pod \"community-operators-jsvv8\" (UID: \"c94b4ce9-b4a7-48d6-940c-198c0e29f213\") " pod="openshift-marketplace/community-operators-jsvv8" Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.700369 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdzfb\" (UniqueName: \"kubernetes.io/projected/c94b4ce9-b4a7-48d6-940c-198c0e29f213-kube-api-access-qdzfb\") pod \"community-operators-jsvv8\" (UID: \"c94b4ce9-b4a7-48d6-940c-198c0e29f213\") " pod="openshift-marketplace/community-operators-jsvv8" Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.700568 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c94b4ce9-b4a7-48d6-940c-198c0e29f213-utilities\") pod \"community-operators-jsvv8\" (UID: \"c94b4ce9-b4a7-48d6-940c-198c0e29f213\") " pod="openshift-marketplace/community-operators-jsvv8" Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.727353 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3e1d70d4-6ada-47bf-a7f0-086ac6b891a0","Type":"ContainerStarted","Data":"f5d5f43bd282e1415a3c16379d324f79ec40ff80083d96d377ad6e39bb117fd1"} Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.727580 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.728893 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" event={"ID":"9461206a-81ad-4c30-a0fe-18ecf0502570","Type":"ContainerStarted","Data":"b31a72c18b8713fc4fb1693bacb27b2bab5936ec56038c9d249f4c3fc878f2d9"} Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.730953 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86","Type":"ContainerStarted","Data":"a1d96daa6b575224cd53013cf7f91bde6c910782c99de496c67a0b84dccb82fc"} Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.731221 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.761836 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.761812101 podStartE2EDuration="36.761812101s" podCreationTimestamp="2025-11-22 05:14:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:15:17.752563015 +0000 UTC m=+1670.326979562" watchObservedRunningTime="2025-11-22 05:15:17.761812101 +0000 UTC m=+1670.336228638" Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.783979 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.783957818 podStartE2EDuration="37.783957818s" podCreationTimestamp="2025-11-22 05:14:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:15:17.774354983 +0000 UTC m=+1670.348771530" watchObservedRunningTime="2025-11-22 05:15:17.783957818 +0000 UTC m=+1670.358374345" Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.802323 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c94b4ce9-b4a7-48d6-940c-198c0e29f213-catalog-content\") pod \"community-operators-jsvv8\" (UID: \"c94b4ce9-b4a7-48d6-940c-198c0e29f213\") " pod="openshift-marketplace/community-operators-jsvv8" Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.802539 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdzfb\" (UniqueName: \"kubernetes.io/projected/c94b4ce9-b4a7-48d6-940c-198c0e29f213-kube-api-access-qdzfb\") pod \"community-operators-jsvv8\" (UID: \"c94b4ce9-b4a7-48d6-940c-198c0e29f213\") " pod="openshift-marketplace/community-operators-jsvv8" Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.802612 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c94b4ce9-b4a7-48d6-940c-198c0e29f213-utilities\") pod \"community-operators-jsvv8\" (UID: \"c94b4ce9-b4a7-48d6-940c-198c0e29f213\") " pod="openshift-marketplace/community-operators-jsvv8" Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.804031 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c94b4ce9-b4a7-48d6-940c-198c0e29f213-catalog-content\") pod \"community-operators-jsvv8\" (UID: \"c94b4ce9-b4a7-48d6-940c-198c0e29f213\") " pod="openshift-marketplace/community-operators-jsvv8" Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.804905 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c94b4ce9-b4a7-48d6-940c-198c0e29f213-utilities\") pod \"community-operators-jsvv8\" (UID: \"c94b4ce9-b4a7-48d6-940c-198c0e29f213\") " pod="openshift-marketplace/community-operators-jsvv8" Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.828348 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdzfb\" (UniqueName: \"kubernetes.io/projected/c94b4ce9-b4a7-48d6-940c-198c0e29f213-kube-api-access-qdzfb\") pod \"community-operators-jsvv8\" (UID: \"c94b4ce9-b4a7-48d6-940c-198c0e29f213\") " pod="openshift-marketplace/community-operators-jsvv8" Nov 22 05:15:17 crc kubenswrapper[4818]: I1122 05:15:17.979502 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jsvv8" Nov 22 05:15:18 crc kubenswrapper[4818]: I1122 05:15:18.539344 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jsvv8"] Nov 22 05:15:18 crc kubenswrapper[4818]: W1122 05:15:18.555476 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc94b4ce9_b4a7_48d6_940c_198c0e29f213.slice/crio-af5c938a19e5c031d4091ca5ca7807d00d04d85f6bff170e8a280697ee79dc71 WatchSource:0}: Error finding container af5c938a19e5c031d4091ca5ca7807d00d04d85f6bff170e8a280697ee79dc71: Status 404 returned error can't find the container with id af5c938a19e5c031d4091ca5ca7807d00d04d85f6bff170e8a280697ee79dc71 Nov 22 05:15:18 crc kubenswrapper[4818]: I1122 05:15:18.740593 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jsvv8" event={"ID":"c94b4ce9-b4a7-48d6-940c-198c0e29f213","Type":"ContainerStarted","Data":"af5c938a19e5c031d4091ca5ca7807d00d04d85f6bff170e8a280697ee79dc71"} Nov 22 05:15:19 crc kubenswrapper[4818]: I1122 05:15:19.750937 4818 generic.go:334] "Generic (PLEG): container finished" podID="c94b4ce9-b4a7-48d6-940c-198c0e29f213" containerID="f33b50a9982a22459fe893391954c5d8f4b26619bc61001be520c676970ba7e0" exitCode=0 Nov 22 05:15:19 crc kubenswrapper[4818]: I1122 05:15:19.751157 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jsvv8" event={"ID":"c94b4ce9-b4a7-48d6-940c-198c0e29f213","Type":"ContainerDied","Data":"f33b50a9982a22459fe893391954c5d8f4b26619bc61001be520c676970ba7e0"} Nov 22 05:15:22 crc kubenswrapper[4818]: I1122 05:15:22.674536 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xhdb2"] Nov 22 05:15:22 crc kubenswrapper[4818]: I1122 05:15:22.676715 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xhdb2" Nov 22 05:15:22 crc kubenswrapper[4818]: I1122 05:15:22.688166 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xhdb2"] Nov 22 05:15:22 crc kubenswrapper[4818]: I1122 05:15:22.700024 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfr4d\" (UniqueName: \"kubernetes.io/projected/6ab8451b-0868-4e2f-bd25-fe117c5ed373-kube-api-access-bfr4d\") pod \"certified-operators-xhdb2\" (UID: \"6ab8451b-0868-4e2f-bd25-fe117c5ed373\") " pod="openshift-marketplace/certified-operators-xhdb2" Nov 22 05:15:22 crc kubenswrapper[4818]: I1122 05:15:22.700093 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ab8451b-0868-4e2f-bd25-fe117c5ed373-catalog-content\") pod \"certified-operators-xhdb2\" (UID: \"6ab8451b-0868-4e2f-bd25-fe117c5ed373\") " pod="openshift-marketplace/certified-operators-xhdb2" Nov 22 05:15:22 crc kubenswrapper[4818]: I1122 05:15:22.700118 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ab8451b-0868-4e2f-bd25-fe117c5ed373-utilities\") pod \"certified-operators-xhdb2\" (UID: \"6ab8451b-0868-4e2f-bd25-fe117c5ed373\") " pod="openshift-marketplace/certified-operators-xhdb2" Nov 22 05:15:22 crc kubenswrapper[4818]: I1122 05:15:22.802382 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfr4d\" (UniqueName: \"kubernetes.io/projected/6ab8451b-0868-4e2f-bd25-fe117c5ed373-kube-api-access-bfr4d\") pod \"certified-operators-xhdb2\" (UID: \"6ab8451b-0868-4e2f-bd25-fe117c5ed373\") " pod="openshift-marketplace/certified-operators-xhdb2" Nov 22 05:15:22 crc kubenswrapper[4818]: I1122 05:15:22.802483 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ab8451b-0868-4e2f-bd25-fe117c5ed373-catalog-content\") pod \"certified-operators-xhdb2\" (UID: \"6ab8451b-0868-4e2f-bd25-fe117c5ed373\") " pod="openshift-marketplace/certified-operators-xhdb2" Nov 22 05:15:22 crc kubenswrapper[4818]: I1122 05:15:22.802512 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ab8451b-0868-4e2f-bd25-fe117c5ed373-utilities\") pod \"certified-operators-xhdb2\" (UID: \"6ab8451b-0868-4e2f-bd25-fe117c5ed373\") " pod="openshift-marketplace/certified-operators-xhdb2" Nov 22 05:15:22 crc kubenswrapper[4818]: I1122 05:15:22.803189 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ab8451b-0868-4e2f-bd25-fe117c5ed373-utilities\") pod \"certified-operators-xhdb2\" (UID: \"6ab8451b-0868-4e2f-bd25-fe117c5ed373\") " pod="openshift-marketplace/certified-operators-xhdb2" Nov 22 05:15:22 crc kubenswrapper[4818]: I1122 05:15:22.803223 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ab8451b-0868-4e2f-bd25-fe117c5ed373-catalog-content\") pod \"certified-operators-xhdb2\" (UID: \"6ab8451b-0868-4e2f-bd25-fe117c5ed373\") " pod="openshift-marketplace/certified-operators-xhdb2" Nov 22 05:15:22 crc kubenswrapper[4818]: I1122 05:15:22.822269 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfr4d\" (UniqueName: \"kubernetes.io/projected/6ab8451b-0868-4e2f-bd25-fe117c5ed373-kube-api-access-bfr4d\") pod \"certified-operators-xhdb2\" (UID: \"6ab8451b-0868-4e2f-bd25-fe117c5ed373\") " pod="openshift-marketplace/certified-operators-xhdb2" Nov 22 05:15:23 crc kubenswrapper[4818]: I1122 05:15:23.001487 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xhdb2" Nov 22 05:15:26 crc kubenswrapper[4818]: I1122 05:15:26.817916 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jsvv8" event={"ID":"c94b4ce9-b4a7-48d6-940c-198c0e29f213","Type":"ContainerStarted","Data":"cfa50e600ca485284ddd8eae0037ab5fd91da694515b2accdccf937ecdbe5aa6"} Nov 22 05:15:26 crc kubenswrapper[4818]: I1122 05:15:26.819906 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" event={"ID":"9461206a-81ad-4c30-a0fe-18ecf0502570","Type":"ContainerStarted","Data":"e2358be6b51f69e9623a757a076795204004cc0647800af27c0c6b26f0544379"} Nov 22 05:15:26 crc kubenswrapper[4818]: I1122 05:15:26.859402 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" podStartSLOduration=1.628665057 podStartE2EDuration="10.859375453s" podCreationTimestamp="2025-11-22 05:15:16 +0000 UTC" firstStartedPulling="2025-11-22 05:15:17.298954638 +0000 UTC m=+1669.873371165" lastFinishedPulling="2025-11-22 05:15:26.529665024 +0000 UTC m=+1679.104081561" observedRunningTime="2025-11-22 05:15:26.858413508 +0000 UTC m=+1679.432830065" watchObservedRunningTime="2025-11-22 05:15:26.859375453 +0000 UTC m=+1679.433791980" Nov 22 05:15:26 crc kubenswrapper[4818]: W1122 05:15:26.935428 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ab8451b_0868_4e2f_bd25_fe117c5ed373.slice/crio-0fb638790c43b0336c463f84537753aec7f1e01f25da9265493e44fe78b46c40 WatchSource:0}: Error finding container 0fb638790c43b0336c463f84537753aec7f1e01f25da9265493e44fe78b46c40: Status 404 returned error can't find the container with id 0fb638790c43b0336c463f84537753aec7f1e01f25da9265493e44fe78b46c40 Nov 22 05:15:26 crc kubenswrapper[4818]: I1122 05:15:26.938135 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xhdb2"] Nov 22 05:15:27 crc kubenswrapper[4818]: I1122 05:15:27.832455 4818 generic.go:334] "Generic (PLEG): container finished" podID="6ab8451b-0868-4e2f-bd25-fe117c5ed373" containerID="04b16ce6cbd81f93a9e7899b532631b5f8a701287d5956e580e97f93ad9b8d0f" exitCode=0 Nov 22 05:15:27 crc kubenswrapper[4818]: I1122 05:15:27.832532 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xhdb2" event={"ID":"6ab8451b-0868-4e2f-bd25-fe117c5ed373","Type":"ContainerDied","Data":"04b16ce6cbd81f93a9e7899b532631b5f8a701287d5956e580e97f93ad9b8d0f"} Nov 22 05:15:27 crc kubenswrapper[4818]: I1122 05:15:27.833028 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xhdb2" event={"ID":"6ab8451b-0868-4e2f-bd25-fe117c5ed373","Type":"ContainerStarted","Data":"0fb638790c43b0336c463f84537753aec7f1e01f25da9265493e44fe78b46c40"} Nov 22 05:15:27 crc kubenswrapper[4818]: I1122 05:15:27.835200 4818 generic.go:334] "Generic (PLEG): container finished" podID="c94b4ce9-b4a7-48d6-940c-198c0e29f213" containerID="cfa50e600ca485284ddd8eae0037ab5fd91da694515b2accdccf937ecdbe5aa6" exitCode=0 Nov 22 05:15:27 crc kubenswrapper[4818]: I1122 05:15:27.835319 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jsvv8" event={"ID":"c94b4ce9-b4a7-48d6-940c-198c0e29f213","Type":"ContainerDied","Data":"cfa50e600ca485284ddd8eae0037ab5fd91da694515b2accdccf937ecdbe5aa6"} Nov 22 05:15:29 crc kubenswrapper[4818]: I1122 05:15:29.858185 4818 generic.go:334] "Generic (PLEG): container finished" podID="6ab8451b-0868-4e2f-bd25-fe117c5ed373" containerID="34b42698e6b8ca5d506bc02b75f2fd2ae9ff93fe69478a871a2b8f7c0e72c062" exitCode=0 Nov 22 05:15:29 crc kubenswrapper[4818]: I1122 05:15:29.858372 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xhdb2" event={"ID":"6ab8451b-0868-4e2f-bd25-fe117c5ed373","Type":"ContainerDied","Data":"34b42698e6b8ca5d506bc02b75f2fd2ae9ff93fe69478a871a2b8f7c0e72c062"} Nov 22 05:15:29 crc kubenswrapper[4818]: I1122 05:15:29.863641 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jsvv8" event={"ID":"c94b4ce9-b4a7-48d6-940c-198c0e29f213","Type":"ContainerStarted","Data":"97aef58e1251d7cfd082e92933e26b93d15df439c4e266afbaec636296c8dd05"} Nov 22 05:15:29 crc kubenswrapper[4818]: I1122 05:15:29.913515 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jsvv8" podStartSLOduration=4.457090388 podStartE2EDuration="12.913418276s" podCreationTimestamp="2025-11-22 05:15:17 +0000 UTC" firstStartedPulling="2025-11-22 05:15:20.20151123 +0000 UTC m=+1672.775927757" lastFinishedPulling="2025-11-22 05:15:28.657839108 +0000 UTC m=+1681.232255645" observedRunningTime="2025-11-22 05:15:29.903467741 +0000 UTC m=+1682.477884278" watchObservedRunningTime="2025-11-22 05:15:29.913418276 +0000 UTC m=+1682.487834813" Nov 22 05:15:30 crc kubenswrapper[4818]: I1122 05:15:30.291798 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:15:30 crc kubenswrapper[4818]: E1122 05:15:30.292060 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:15:30 crc kubenswrapper[4818]: I1122 05:15:30.873662 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xhdb2" event={"ID":"6ab8451b-0868-4e2f-bd25-fe117c5ed373","Type":"ContainerStarted","Data":"1bf6897ff9956ca70a4fda93f1b6dfd8812a2d9102d51dc5a56a39fdfde98c25"} Nov 22 05:15:30 crc kubenswrapper[4818]: I1122 05:15:30.898527 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xhdb2" podStartSLOduration=6.435041765 podStartE2EDuration="8.898508756s" podCreationTimestamp="2025-11-22 05:15:22 +0000 UTC" firstStartedPulling="2025-11-22 05:15:27.836080801 +0000 UTC m=+1680.410497328" lastFinishedPulling="2025-11-22 05:15:30.299547792 +0000 UTC m=+1682.873964319" observedRunningTime="2025-11-22 05:15:30.894750096 +0000 UTC m=+1683.469166633" watchObservedRunningTime="2025-11-22 05:15:30.898508756 +0000 UTC m=+1683.472925283" Nov 22 05:15:31 crc kubenswrapper[4818]: I1122 05:15:31.008234 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 22 05:15:31 crc kubenswrapper[4818]: I1122 05:15:31.685459 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 22 05:15:33 crc kubenswrapper[4818]: I1122 05:15:33.002225 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xhdb2" Nov 22 05:15:33 crc kubenswrapper[4818]: I1122 05:15:33.002589 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xhdb2" Nov 22 05:15:33 crc kubenswrapper[4818]: I1122 05:15:33.053058 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xhdb2" Nov 22 05:15:37 crc kubenswrapper[4818]: I1122 05:15:37.980122 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jsvv8" Nov 22 05:15:37 crc kubenswrapper[4818]: I1122 05:15:37.980702 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jsvv8" Nov 22 05:15:38 crc kubenswrapper[4818]: I1122 05:15:38.025803 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jsvv8" Nov 22 05:15:39 crc kubenswrapper[4818]: I1122 05:15:39.048923 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jsvv8" Nov 22 05:15:39 crc kubenswrapper[4818]: I1122 05:15:39.109993 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jsvv8"] Nov 22 05:15:39 crc kubenswrapper[4818]: I1122 05:15:39.987018 4818 generic.go:334] "Generic (PLEG): container finished" podID="9461206a-81ad-4c30-a0fe-18ecf0502570" containerID="e2358be6b51f69e9623a757a076795204004cc0647800af27c0c6b26f0544379" exitCode=0 Nov 22 05:15:39 crc kubenswrapper[4818]: I1122 05:15:39.987086 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" event={"ID":"9461206a-81ad-4c30-a0fe-18ecf0502570","Type":"ContainerDied","Data":"e2358be6b51f69e9623a757a076795204004cc0647800af27c0c6b26f0544379"} Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:40.997639 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jsvv8" podUID="c94b4ce9-b4a7-48d6-940c-198c0e29f213" containerName="registry-server" containerID="cri-o://97aef58e1251d7cfd082e92933e26b93d15df439c4e266afbaec636296c8dd05" gracePeriod=2 Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.517839 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.525243 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jsvv8" Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.603192 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9461206a-81ad-4c30-a0fe-18ecf0502570-repo-setup-combined-ca-bundle\") pod \"9461206a-81ad-4c30-a0fe-18ecf0502570\" (UID: \"9461206a-81ad-4c30-a0fe-18ecf0502570\") " Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.603284 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c94b4ce9-b4a7-48d6-940c-198c0e29f213-catalog-content\") pod \"c94b4ce9-b4a7-48d6-940c-198c0e29f213\" (UID: \"c94b4ce9-b4a7-48d6-940c-198c0e29f213\") " Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.603422 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9461206a-81ad-4c30-a0fe-18ecf0502570-ssh-key\") pod \"9461206a-81ad-4c30-a0fe-18ecf0502570\" (UID: \"9461206a-81ad-4c30-a0fe-18ecf0502570\") " Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.603453 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c94b4ce9-b4a7-48d6-940c-198c0e29f213-utilities\") pod \"c94b4ce9-b4a7-48d6-940c-198c0e29f213\" (UID: \"c94b4ce9-b4a7-48d6-940c-198c0e29f213\") " Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.603482 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9461206a-81ad-4c30-a0fe-18ecf0502570-inventory\") pod \"9461206a-81ad-4c30-a0fe-18ecf0502570\" (UID: \"9461206a-81ad-4c30-a0fe-18ecf0502570\") " Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.603533 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdzfb\" (UniqueName: \"kubernetes.io/projected/c94b4ce9-b4a7-48d6-940c-198c0e29f213-kube-api-access-qdzfb\") pod \"c94b4ce9-b4a7-48d6-940c-198c0e29f213\" (UID: \"c94b4ce9-b4a7-48d6-940c-198c0e29f213\") " Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.603550 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59w7v\" (UniqueName: \"kubernetes.io/projected/9461206a-81ad-4c30-a0fe-18ecf0502570-kube-api-access-59w7v\") pod \"9461206a-81ad-4c30-a0fe-18ecf0502570\" (UID: \"9461206a-81ad-4c30-a0fe-18ecf0502570\") " Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.605267 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c94b4ce9-b4a7-48d6-940c-198c0e29f213-utilities" (OuterVolumeSpecName: "utilities") pod "c94b4ce9-b4a7-48d6-940c-198c0e29f213" (UID: "c94b4ce9-b4a7-48d6-940c-198c0e29f213"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.621176 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9461206a-81ad-4c30-a0fe-18ecf0502570-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "9461206a-81ad-4c30-a0fe-18ecf0502570" (UID: "9461206a-81ad-4c30-a0fe-18ecf0502570"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.621185 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c94b4ce9-b4a7-48d6-940c-198c0e29f213-kube-api-access-qdzfb" (OuterVolumeSpecName: "kube-api-access-qdzfb") pod "c94b4ce9-b4a7-48d6-940c-198c0e29f213" (UID: "c94b4ce9-b4a7-48d6-940c-198c0e29f213"). InnerVolumeSpecName "kube-api-access-qdzfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.621467 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9461206a-81ad-4c30-a0fe-18ecf0502570-kube-api-access-59w7v" (OuterVolumeSpecName: "kube-api-access-59w7v") pod "9461206a-81ad-4c30-a0fe-18ecf0502570" (UID: "9461206a-81ad-4c30-a0fe-18ecf0502570"). InnerVolumeSpecName "kube-api-access-59w7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.632739 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9461206a-81ad-4c30-a0fe-18ecf0502570-inventory" (OuterVolumeSpecName: "inventory") pod "9461206a-81ad-4c30-a0fe-18ecf0502570" (UID: "9461206a-81ad-4c30-a0fe-18ecf0502570"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.634851 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9461206a-81ad-4c30-a0fe-18ecf0502570-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9461206a-81ad-4c30-a0fe-18ecf0502570" (UID: "9461206a-81ad-4c30-a0fe-18ecf0502570"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.660790 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c94b4ce9-b4a7-48d6-940c-198c0e29f213-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c94b4ce9-b4a7-48d6-940c-198c0e29f213" (UID: "c94b4ce9-b4a7-48d6-940c-198c0e29f213"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.704782 4818 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9461206a-81ad-4c30-a0fe-18ecf0502570-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.704810 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c94b4ce9-b4a7-48d6-940c-198c0e29f213-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.704821 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9461206a-81ad-4c30-a0fe-18ecf0502570-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.704831 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c94b4ce9-b4a7-48d6-940c-198c0e29f213-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.704841 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9461206a-81ad-4c30-a0fe-18ecf0502570-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.704849 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdzfb\" (UniqueName: \"kubernetes.io/projected/c94b4ce9-b4a7-48d6-940c-198c0e29f213-kube-api-access-qdzfb\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:41 crc kubenswrapper[4818]: I1122 05:15:41.704858 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59w7v\" (UniqueName: \"kubernetes.io/projected/9461206a-81ad-4c30-a0fe-18ecf0502570-kube-api-access-59w7v\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.012522 4818 generic.go:334] "Generic (PLEG): container finished" podID="c94b4ce9-b4a7-48d6-940c-198c0e29f213" containerID="97aef58e1251d7cfd082e92933e26b93d15df439c4e266afbaec636296c8dd05" exitCode=0 Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.012639 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jsvv8" event={"ID":"c94b4ce9-b4a7-48d6-940c-198c0e29f213","Type":"ContainerDied","Data":"97aef58e1251d7cfd082e92933e26b93d15df439c4e266afbaec636296c8dd05"} Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.012948 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jsvv8" event={"ID":"c94b4ce9-b4a7-48d6-940c-198c0e29f213","Type":"ContainerDied","Data":"af5c938a19e5c031d4091ca5ca7807d00d04d85f6bff170e8a280697ee79dc71"} Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.012976 4818 scope.go:117] "RemoveContainer" containerID="97aef58e1251d7cfd082e92933e26b93d15df439c4e266afbaec636296c8dd05" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.012692 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jsvv8" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.016318 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" event={"ID":"9461206a-81ad-4c30-a0fe-18ecf0502570","Type":"ContainerDied","Data":"b31a72c18b8713fc4fb1693bacb27b2bab5936ec56038c9d249f4c3fc878f2d9"} Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.016363 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b31a72c18b8713fc4fb1693bacb27b2bab5936ec56038c9d249f4c3fc878f2d9" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.016454 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.074423 4818 scope.go:117] "RemoveContainer" containerID="cfa50e600ca485284ddd8eae0037ab5fd91da694515b2accdccf937ecdbe5aa6" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.082869 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jsvv8"] Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.094717 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw"] Nov 22 05:15:42 crc kubenswrapper[4818]: E1122 05:15:42.095471 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c94b4ce9-b4a7-48d6-940c-198c0e29f213" containerName="extract-content" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.095518 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c94b4ce9-b4a7-48d6-940c-198c0e29f213" containerName="extract-content" Nov 22 05:15:42 crc kubenswrapper[4818]: E1122 05:15:42.095564 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c94b4ce9-b4a7-48d6-940c-198c0e29f213" containerName="extract-utilities" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.095582 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c94b4ce9-b4a7-48d6-940c-198c0e29f213" containerName="extract-utilities" Nov 22 05:15:42 crc kubenswrapper[4818]: E1122 05:15:42.095608 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c94b4ce9-b4a7-48d6-940c-198c0e29f213" containerName="registry-server" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.095623 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c94b4ce9-b4a7-48d6-940c-198c0e29f213" containerName="registry-server" Nov 22 05:15:42 crc kubenswrapper[4818]: E1122 05:15:42.095706 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9461206a-81ad-4c30-a0fe-18ecf0502570" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.095725 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="9461206a-81ad-4c30-a0fe-18ecf0502570" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.096087 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="9461206a-81ad-4c30-a0fe-18ecf0502570" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.096161 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="c94b4ce9-b4a7-48d6-940c-198c0e29f213" containerName="registry-server" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.097443 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.109477 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jsvv8"] Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.113918 4818 scope.go:117] "RemoveContainer" containerID="f33b50a9982a22459fe893391954c5d8f4b26619bc61001be520c676970ba7e0" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.114137 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.114337 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.114446 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.114639 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.119407 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw"] Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.169744 4818 scope.go:117] "RemoveContainer" containerID="97aef58e1251d7cfd082e92933e26b93d15df439c4e266afbaec636296c8dd05" Nov 22 05:15:42 crc kubenswrapper[4818]: E1122 05:15:42.170420 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97aef58e1251d7cfd082e92933e26b93d15df439c4e266afbaec636296c8dd05\": container with ID starting with 97aef58e1251d7cfd082e92933e26b93d15df439c4e266afbaec636296c8dd05 not found: ID does not exist" containerID="97aef58e1251d7cfd082e92933e26b93d15df439c4e266afbaec636296c8dd05" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.170532 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97aef58e1251d7cfd082e92933e26b93d15df439c4e266afbaec636296c8dd05"} err="failed to get container status \"97aef58e1251d7cfd082e92933e26b93d15df439c4e266afbaec636296c8dd05\": rpc error: code = NotFound desc = could not find container \"97aef58e1251d7cfd082e92933e26b93d15df439c4e266afbaec636296c8dd05\": container with ID starting with 97aef58e1251d7cfd082e92933e26b93d15df439c4e266afbaec636296c8dd05 not found: ID does not exist" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.170680 4818 scope.go:117] "RemoveContainer" containerID="cfa50e600ca485284ddd8eae0037ab5fd91da694515b2accdccf937ecdbe5aa6" Nov 22 05:15:42 crc kubenswrapper[4818]: E1122 05:15:42.171013 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfa50e600ca485284ddd8eae0037ab5fd91da694515b2accdccf937ecdbe5aa6\": container with ID starting with cfa50e600ca485284ddd8eae0037ab5fd91da694515b2accdccf937ecdbe5aa6 not found: ID does not exist" containerID="cfa50e600ca485284ddd8eae0037ab5fd91da694515b2accdccf937ecdbe5aa6" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.171044 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfa50e600ca485284ddd8eae0037ab5fd91da694515b2accdccf937ecdbe5aa6"} err="failed to get container status \"cfa50e600ca485284ddd8eae0037ab5fd91da694515b2accdccf937ecdbe5aa6\": rpc error: code = NotFound desc = could not find container \"cfa50e600ca485284ddd8eae0037ab5fd91da694515b2accdccf937ecdbe5aa6\": container with ID starting with cfa50e600ca485284ddd8eae0037ab5fd91da694515b2accdccf937ecdbe5aa6 not found: ID does not exist" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.171065 4818 scope.go:117] "RemoveContainer" containerID="f33b50a9982a22459fe893391954c5d8f4b26619bc61001be520c676970ba7e0" Nov 22 05:15:42 crc kubenswrapper[4818]: E1122 05:15:42.172114 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f33b50a9982a22459fe893391954c5d8f4b26619bc61001be520c676970ba7e0\": container with ID starting with f33b50a9982a22459fe893391954c5d8f4b26619bc61001be520c676970ba7e0 not found: ID does not exist" containerID="f33b50a9982a22459fe893391954c5d8f4b26619bc61001be520c676970ba7e0" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.172292 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f33b50a9982a22459fe893391954c5d8f4b26619bc61001be520c676970ba7e0"} err="failed to get container status \"f33b50a9982a22459fe893391954c5d8f4b26619bc61001be520c676970ba7e0\": rpc error: code = NotFound desc = could not find container \"f33b50a9982a22459fe893391954c5d8f4b26619bc61001be520c676970ba7e0\": container with ID starting with f33b50a9982a22459fe893391954c5d8f4b26619bc61001be520c676970ba7e0 not found: ID does not exist" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.216094 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw\" (UID: \"2ffc9da3-3fd2-4752-9d8d-f22e149e375d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.216644 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw\" (UID: \"2ffc9da3-3fd2-4752-9d8d-f22e149e375d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.216947 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw\" (UID: \"2ffc9da3-3fd2-4752-9d8d-f22e149e375d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.217167 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw9xs\" (UniqueName: \"kubernetes.io/projected/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-kube-api-access-lw9xs\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw\" (UID: \"2ffc9da3-3fd2-4752-9d8d-f22e149e375d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.310738 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c94b4ce9-b4a7-48d6-940c-198c0e29f213" path="/var/lib/kubelet/pods/c94b4ce9-b4a7-48d6-940c-198c0e29f213/volumes" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.320054 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw\" (UID: \"2ffc9da3-3fd2-4752-9d8d-f22e149e375d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.320446 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw\" (UID: \"2ffc9da3-3fd2-4752-9d8d-f22e149e375d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.320667 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw9xs\" (UniqueName: \"kubernetes.io/projected/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-kube-api-access-lw9xs\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw\" (UID: \"2ffc9da3-3fd2-4752-9d8d-f22e149e375d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.321240 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw\" (UID: \"2ffc9da3-3fd2-4752-9d8d-f22e149e375d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.325929 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw\" (UID: \"2ffc9da3-3fd2-4752-9d8d-f22e149e375d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.327768 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw\" (UID: \"2ffc9da3-3fd2-4752-9d8d-f22e149e375d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.328317 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw\" (UID: \"2ffc9da3-3fd2-4752-9d8d-f22e149e375d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.341238 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw9xs\" (UniqueName: \"kubernetes.io/projected/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-kube-api-access-lw9xs\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw\" (UID: \"2ffc9da3-3fd2-4752-9d8d-f22e149e375d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" Nov 22 05:15:42 crc kubenswrapper[4818]: I1122 05:15:42.495911 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" Nov 22 05:15:43 crc kubenswrapper[4818]: W1122 05:15:43.081973 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ffc9da3_3fd2_4752_9d8d_f22e149e375d.slice/crio-5183fdede762c060dc555fae0fee018a0de9968799cc7f758b999922e29c6757 WatchSource:0}: Error finding container 5183fdede762c060dc555fae0fee018a0de9968799cc7f758b999922e29c6757: Status 404 returned error can't find the container with id 5183fdede762c060dc555fae0fee018a0de9968799cc7f758b999922e29c6757 Nov 22 05:15:43 crc kubenswrapper[4818]: I1122 05:15:43.082036 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw"] Nov 22 05:15:43 crc kubenswrapper[4818]: I1122 05:15:43.100027 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xhdb2" Nov 22 05:15:43 crc kubenswrapper[4818]: I1122 05:15:43.775792 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xhdb2"] Nov 22 05:15:44 crc kubenswrapper[4818]: I1122 05:15:44.046220 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" event={"ID":"2ffc9da3-3fd2-4752-9d8d-f22e149e375d","Type":"ContainerStarted","Data":"342208bbd7228f9f32a772585b9ad640b0713c4700a26825ae4946a4c5276f53"} Nov 22 05:15:44 crc kubenswrapper[4818]: I1122 05:15:44.046315 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" event={"ID":"2ffc9da3-3fd2-4752-9d8d-f22e149e375d","Type":"ContainerStarted","Data":"5183fdede762c060dc555fae0fee018a0de9968799cc7f758b999922e29c6757"} Nov 22 05:15:44 crc kubenswrapper[4818]: I1122 05:15:44.046363 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xhdb2" podUID="6ab8451b-0868-4e2f-bd25-fe117c5ed373" containerName="registry-server" containerID="cri-o://1bf6897ff9956ca70a4fda93f1b6dfd8812a2d9102d51dc5a56a39fdfde98c25" gracePeriod=2 Nov 22 05:15:44 crc kubenswrapper[4818]: I1122 05:15:44.073806 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" podStartSLOduration=1.639638056 podStartE2EDuration="2.073728765s" podCreationTimestamp="2025-11-22 05:15:42 +0000 UTC" firstStartedPulling="2025-11-22 05:15:43.084950617 +0000 UTC m=+1695.659367144" lastFinishedPulling="2025-11-22 05:15:43.519041336 +0000 UTC m=+1696.093457853" observedRunningTime="2025-11-22 05:15:44.070986912 +0000 UTC m=+1696.645403439" watchObservedRunningTime="2025-11-22 05:15:44.073728765 +0000 UTC m=+1696.648184053" Nov 22 05:15:44 crc kubenswrapper[4818]: I1122 05:15:44.505938 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xhdb2" Nov 22 05:15:44 crc kubenswrapper[4818]: I1122 05:15:44.562567 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfr4d\" (UniqueName: \"kubernetes.io/projected/6ab8451b-0868-4e2f-bd25-fe117c5ed373-kube-api-access-bfr4d\") pod \"6ab8451b-0868-4e2f-bd25-fe117c5ed373\" (UID: \"6ab8451b-0868-4e2f-bd25-fe117c5ed373\") " Nov 22 05:15:44 crc kubenswrapper[4818]: I1122 05:15:44.562660 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ab8451b-0868-4e2f-bd25-fe117c5ed373-utilities\") pod \"6ab8451b-0868-4e2f-bd25-fe117c5ed373\" (UID: \"6ab8451b-0868-4e2f-bd25-fe117c5ed373\") " Nov 22 05:15:44 crc kubenswrapper[4818]: I1122 05:15:44.562710 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ab8451b-0868-4e2f-bd25-fe117c5ed373-catalog-content\") pod \"6ab8451b-0868-4e2f-bd25-fe117c5ed373\" (UID: \"6ab8451b-0868-4e2f-bd25-fe117c5ed373\") " Nov 22 05:15:44 crc kubenswrapper[4818]: I1122 05:15:44.569643 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ab8451b-0868-4e2f-bd25-fe117c5ed373-kube-api-access-bfr4d" (OuterVolumeSpecName: "kube-api-access-bfr4d") pod "6ab8451b-0868-4e2f-bd25-fe117c5ed373" (UID: "6ab8451b-0868-4e2f-bd25-fe117c5ed373"). InnerVolumeSpecName "kube-api-access-bfr4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:15:44 crc kubenswrapper[4818]: I1122 05:15:44.577507 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ab8451b-0868-4e2f-bd25-fe117c5ed373-utilities" (OuterVolumeSpecName: "utilities") pod "6ab8451b-0868-4e2f-bd25-fe117c5ed373" (UID: "6ab8451b-0868-4e2f-bd25-fe117c5ed373"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:15:44 crc kubenswrapper[4818]: I1122 05:15:44.624174 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ab8451b-0868-4e2f-bd25-fe117c5ed373-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ab8451b-0868-4e2f-bd25-fe117c5ed373" (UID: "6ab8451b-0868-4e2f-bd25-fe117c5ed373"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:15:44 crc kubenswrapper[4818]: I1122 05:15:44.664840 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfr4d\" (UniqueName: \"kubernetes.io/projected/6ab8451b-0868-4e2f-bd25-fe117c5ed373-kube-api-access-bfr4d\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:44 crc kubenswrapper[4818]: I1122 05:15:44.664873 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ab8451b-0868-4e2f-bd25-fe117c5ed373-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:44 crc kubenswrapper[4818]: I1122 05:15:44.664883 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ab8451b-0868-4e2f-bd25-fe117c5ed373-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 05:15:45 crc kubenswrapper[4818]: I1122 05:15:45.063133 4818 generic.go:334] "Generic (PLEG): container finished" podID="6ab8451b-0868-4e2f-bd25-fe117c5ed373" containerID="1bf6897ff9956ca70a4fda93f1b6dfd8812a2d9102d51dc5a56a39fdfde98c25" exitCode=0 Nov 22 05:15:45 crc kubenswrapper[4818]: I1122 05:15:45.063216 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xhdb2" Nov 22 05:15:45 crc kubenswrapper[4818]: I1122 05:15:45.063210 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xhdb2" event={"ID":"6ab8451b-0868-4e2f-bd25-fe117c5ed373","Type":"ContainerDied","Data":"1bf6897ff9956ca70a4fda93f1b6dfd8812a2d9102d51dc5a56a39fdfde98c25"} Nov 22 05:15:45 crc kubenswrapper[4818]: I1122 05:15:45.063794 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xhdb2" event={"ID":"6ab8451b-0868-4e2f-bd25-fe117c5ed373","Type":"ContainerDied","Data":"0fb638790c43b0336c463f84537753aec7f1e01f25da9265493e44fe78b46c40"} Nov 22 05:15:45 crc kubenswrapper[4818]: I1122 05:15:45.063829 4818 scope.go:117] "RemoveContainer" containerID="1bf6897ff9956ca70a4fda93f1b6dfd8812a2d9102d51dc5a56a39fdfde98c25" Nov 22 05:15:45 crc kubenswrapper[4818]: I1122 05:15:45.107048 4818 scope.go:117] "RemoveContainer" containerID="34b42698e6b8ca5d506bc02b75f2fd2ae9ff93fe69478a871a2b8f7c0e72c062" Nov 22 05:15:45 crc kubenswrapper[4818]: I1122 05:15:45.117156 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xhdb2"] Nov 22 05:15:45 crc kubenswrapper[4818]: I1122 05:15:45.125894 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xhdb2"] Nov 22 05:15:45 crc kubenswrapper[4818]: I1122 05:15:45.144878 4818 scope.go:117] "RemoveContainer" containerID="04b16ce6cbd81f93a9e7899b532631b5f8a701287d5956e580e97f93ad9b8d0f" Nov 22 05:15:45 crc kubenswrapper[4818]: I1122 05:15:45.200433 4818 scope.go:117] "RemoveContainer" containerID="1bf6897ff9956ca70a4fda93f1b6dfd8812a2d9102d51dc5a56a39fdfde98c25" Nov 22 05:15:45 crc kubenswrapper[4818]: E1122 05:15:45.200907 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bf6897ff9956ca70a4fda93f1b6dfd8812a2d9102d51dc5a56a39fdfde98c25\": container with ID starting with 1bf6897ff9956ca70a4fda93f1b6dfd8812a2d9102d51dc5a56a39fdfde98c25 not found: ID does not exist" containerID="1bf6897ff9956ca70a4fda93f1b6dfd8812a2d9102d51dc5a56a39fdfde98c25" Nov 22 05:15:45 crc kubenswrapper[4818]: I1122 05:15:45.200963 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bf6897ff9956ca70a4fda93f1b6dfd8812a2d9102d51dc5a56a39fdfde98c25"} err="failed to get container status \"1bf6897ff9956ca70a4fda93f1b6dfd8812a2d9102d51dc5a56a39fdfde98c25\": rpc error: code = NotFound desc = could not find container \"1bf6897ff9956ca70a4fda93f1b6dfd8812a2d9102d51dc5a56a39fdfde98c25\": container with ID starting with 1bf6897ff9956ca70a4fda93f1b6dfd8812a2d9102d51dc5a56a39fdfde98c25 not found: ID does not exist" Nov 22 05:15:45 crc kubenswrapper[4818]: I1122 05:15:45.200996 4818 scope.go:117] "RemoveContainer" containerID="34b42698e6b8ca5d506bc02b75f2fd2ae9ff93fe69478a871a2b8f7c0e72c062" Nov 22 05:15:45 crc kubenswrapper[4818]: E1122 05:15:45.201409 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34b42698e6b8ca5d506bc02b75f2fd2ae9ff93fe69478a871a2b8f7c0e72c062\": container with ID starting with 34b42698e6b8ca5d506bc02b75f2fd2ae9ff93fe69478a871a2b8f7c0e72c062 not found: ID does not exist" containerID="34b42698e6b8ca5d506bc02b75f2fd2ae9ff93fe69478a871a2b8f7c0e72c062" Nov 22 05:15:45 crc kubenswrapper[4818]: I1122 05:15:45.201456 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34b42698e6b8ca5d506bc02b75f2fd2ae9ff93fe69478a871a2b8f7c0e72c062"} err="failed to get container status \"34b42698e6b8ca5d506bc02b75f2fd2ae9ff93fe69478a871a2b8f7c0e72c062\": rpc error: code = NotFound desc = could not find container \"34b42698e6b8ca5d506bc02b75f2fd2ae9ff93fe69478a871a2b8f7c0e72c062\": container with ID starting with 34b42698e6b8ca5d506bc02b75f2fd2ae9ff93fe69478a871a2b8f7c0e72c062 not found: ID does not exist" Nov 22 05:15:45 crc kubenswrapper[4818]: I1122 05:15:45.201491 4818 scope.go:117] "RemoveContainer" containerID="04b16ce6cbd81f93a9e7899b532631b5f8a701287d5956e580e97f93ad9b8d0f" Nov 22 05:15:45 crc kubenswrapper[4818]: E1122 05:15:45.201788 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04b16ce6cbd81f93a9e7899b532631b5f8a701287d5956e580e97f93ad9b8d0f\": container with ID starting with 04b16ce6cbd81f93a9e7899b532631b5f8a701287d5956e580e97f93ad9b8d0f not found: ID does not exist" containerID="04b16ce6cbd81f93a9e7899b532631b5f8a701287d5956e580e97f93ad9b8d0f" Nov 22 05:15:45 crc kubenswrapper[4818]: I1122 05:15:45.201817 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04b16ce6cbd81f93a9e7899b532631b5f8a701287d5956e580e97f93ad9b8d0f"} err="failed to get container status \"04b16ce6cbd81f93a9e7899b532631b5f8a701287d5956e580e97f93ad9b8d0f\": rpc error: code = NotFound desc = could not find container \"04b16ce6cbd81f93a9e7899b532631b5f8a701287d5956e580e97f93ad9b8d0f\": container with ID starting with 04b16ce6cbd81f93a9e7899b532631b5f8a701287d5956e580e97f93ad9b8d0f not found: ID does not exist" Nov 22 05:15:45 crc kubenswrapper[4818]: I1122 05:15:45.290979 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:15:45 crc kubenswrapper[4818]: E1122 05:15:45.291217 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:15:46 crc kubenswrapper[4818]: I1122 05:15:46.308841 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ab8451b-0868-4e2f-bd25-fe117c5ed373" path="/var/lib/kubelet/pods/6ab8451b-0868-4e2f-bd25-fe117c5ed373/volumes" Nov 22 05:16:00 crc kubenswrapper[4818]: I1122 05:16:00.291441 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:16:00 crc kubenswrapper[4818]: E1122 05:16:00.292607 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:16:15 crc kubenswrapper[4818]: I1122 05:16:15.291181 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:16:15 crc kubenswrapper[4818]: E1122 05:16:15.292071 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:16:30 crc kubenswrapper[4818]: I1122 05:16:30.292390 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:16:30 crc kubenswrapper[4818]: E1122 05:16:30.293608 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:16:37 crc kubenswrapper[4818]: I1122 05:16:37.122434 4818 scope.go:117] "RemoveContainer" containerID="d487a8c5d84ba9f8f05636aff782142b966c55d4f278ab53dec50fb2b1a1a854" Nov 22 05:16:37 crc kubenswrapper[4818]: I1122 05:16:37.148587 4818 scope.go:117] "RemoveContainer" containerID="3cba6a520a9001c55a8f1e00a44043e22e6e8f1ab4eb4d7d10157a2d09ab9561" Nov 22 05:16:41 crc kubenswrapper[4818]: I1122 05:16:41.291639 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:16:41 crc kubenswrapper[4818]: E1122 05:16:41.292331 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:16:52 crc kubenswrapper[4818]: I1122 05:16:52.293733 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:16:52 crc kubenswrapper[4818]: E1122 05:16:52.294393 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:17:04 crc kubenswrapper[4818]: I1122 05:17:04.291786 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:17:04 crc kubenswrapper[4818]: E1122 05:17:04.296966 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:17:16 crc kubenswrapper[4818]: I1122 05:17:16.291974 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:17:16 crc kubenswrapper[4818]: E1122 05:17:16.292907 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:17:29 crc kubenswrapper[4818]: I1122 05:17:29.291345 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:17:29 crc kubenswrapper[4818]: E1122 05:17:29.292203 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:17:37 crc kubenswrapper[4818]: I1122 05:17:37.271962 4818 scope.go:117] "RemoveContainer" containerID="7cb97b13bbe6cb54623a632b4f7c9b161afb6c86f35e9e141fd7166dc28a44e2" Nov 22 05:17:37 crc kubenswrapper[4818]: I1122 05:17:37.295858 4818 scope.go:117] "RemoveContainer" containerID="3c7694200619b732d5d71a5c66213d99c4dc98ba443c3669d13766ec94e389b6" Nov 22 05:17:37 crc kubenswrapper[4818]: I1122 05:17:37.321500 4818 scope.go:117] "RemoveContainer" containerID="f7a4d0ea11175f50c5e59c18c7fb82d32bcd64c49e2e4b60395c59ecb1aa066d" Nov 22 05:17:37 crc kubenswrapper[4818]: I1122 05:17:37.358511 4818 scope.go:117] "RemoveContainer" containerID="86b9735df77b1ddcf2089199f40768b0089f2085555bfff643165c859e3a44bc" Nov 22 05:17:44 crc kubenswrapper[4818]: I1122 05:17:44.291445 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:17:44 crc kubenswrapper[4818]: E1122 05:17:44.292784 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:17:58 crc kubenswrapper[4818]: I1122 05:17:58.303384 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:17:58 crc kubenswrapper[4818]: E1122 05:17:58.304336 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:18:12 crc kubenswrapper[4818]: I1122 05:18:12.292183 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:18:12 crc kubenswrapper[4818]: E1122 05:18:12.293024 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:18:23 crc kubenswrapper[4818]: I1122 05:18:23.292196 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:18:23 crc kubenswrapper[4818]: E1122 05:18:23.293401 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:18:36 crc kubenswrapper[4818]: I1122 05:18:36.292022 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:18:36 crc kubenswrapper[4818]: E1122 05:18:36.293359 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:18:51 crc kubenswrapper[4818]: I1122 05:18:51.291615 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:18:51 crc kubenswrapper[4818]: E1122 05:18:51.293985 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:18:54 crc kubenswrapper[4818]: I1122 05:18:54.021186 4818 generic.go:334] "Generic (PLEG): container finished" podID="2ffc9da3-3fd2-4752-9d8d-f22e149e375d" containerID="342208bbd7228f9f32a772585b9ad640b0713c4700a26825ae4946a4c5276f53" exitCode=0 Nov 22 05:18:54 crc kubenswrapper[4818]: I1122 05:18:54.021945 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" event={"ID":"2ffc9da3-3fd2-4752-9d8d-f22e149e375d","Type":"ContainerDied","Data":"342208bbd7228f9f32a772585b9ad640b0713c4700a26825ae4946a4c5276f53"} Nov 22 05:18:55 crc kubenswrapper[4818]: I1122 05:18:55.480233 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" Nov 22 05:18:55 crc kubenswrapper[4818]: I1122 05:18:55.595694 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lw9xs\" (UniqueName: \"kubernetes.io/projected/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-kube-api-access-lw9xs\") pod \"2ffc9da3-3fd2-4752-9d8d-f22e149e375d\" (UID: \"2ffc9da3-3fd2-4752-9d8d-f22e149e375d\") " Nov 22 05:18:55 crc kubenswrapper[4818]: I1122 05:18:55.595774 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-ssh-key\") pod \"2ffc9da3-3fd2-4752-9d8d-f22e149e375d\" (UID: \"2ffc9da3-3fd2-4752-9d8d-f22e149e375d\") " Nov 22 05:18:55 crc kubenswrapper[4818]: I1122 05:18:55.595852 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-inventory\") pod \"2ffc9da3-3fd2-4752-9d8d-f22e149e375d\" (UID: \"2ffc9da3-3fd2-4752-9d8d-f22e149e375d\") " Nov 22 05:18:55 crc kubenswrapper[4818]: I1122 05:18:55.595957 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-bootstrap-combined-ca-bundle\") pod \"2ffc9da3-3fd2-4752-9d8d-f22e149e375d\" (UID: \"2ffc9da3-3fd2-4752-9d8d-f22e149e375d\") " Nov 22 05:18:55 crc kubenswrapper[4818]: I1122 05:18:55.604485 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "2ffc9da3-3fd2-4752-9d8d-f22e149e375d" (UID: "2ffc9da3-3fd2-4752-9d8d-f22e149e375d"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:18:55 crc kubenswrapper[4818]: I1122 05:18:55.604507 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-kube-api-access-lw9xs" (OuterVolumeSpecName: "kube-api-access-lw9xs") pod "2ffc9da3-3fd2-4752-9d8d-f22e149e375d" (UID: "2ffc9da3-3fd2-4752-9d8d-f22e149e375d"). InnerVolumeSpecName "kube-api-access-lw9xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:18:55 crc kubenswrapper[4818]: I1122 05:18:55.623014 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-inventory" (OuterVolumeSpecName: "inventory") pod "2ffc9da3-3fd2-4752-9d8d-f22e149e375d" (UID: "2ffc9da3-3fd2-4752-9d8d-f22e149e375d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:18:55 crc kubenswrapper[4818]: I1122 05:18:55.633390 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2ffc9da3-3fd2-4752-9d8d-f22e149e375d" (UID: "2ffc9da3-3fd2-4752-9d8d-f22e149e375d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:18:55 crc kubenswrapper[4818]: I1122 05:18:55.697685 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:18:55 crc kubenswrapper[4818]: I1122 05:18:55.697737 4818 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:18:55 crc kubenswrapper[4818]: I1122 05:18:55.697748 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lw9xs\" (UniqueName: \"kubernetes.io/projected/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-kube-api-access-lw9xs\") on node \"crc\" DevicePath \"\"" Nov 22 05:18:55 crc kubenswrapper[4818]: I1122 05:18:55.697759 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ffc9da3-3fd2-4752-9d8d-f22e149e375d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.047353 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" event={"ID":"2ffc9da3-3fd2-4752-9d8d-f22e149e375d","Type":"ContainerDied","Data":"5183fdede762c060dc555fae0fee018a0de9968799cc7f758b999922e29c6757"} Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.047413 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5183fdede762c060dc555fae0fee018a0de9968799cc7f758b999922e29c6757" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.047441 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.119201 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv"] Nov 22 05:18:56 crc kubenswrapper[4818]: E1122 05:18:56.119551 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ab8451b-0868-4e2f-bd25-fe117c5ed373" containerName="extract-utilities" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.119566 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ab8451b-0868-4e2f-bd25-fe117c5ed373" containerName="extract-utilities" Nov 22 05:18:56 crc kubenswrapper[4818]: E1122 05:18:56.119797 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ffc9da3-3fd2-4752-9d8d-f22e149e375d" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.119805 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ffc9da3-3fd2-4752-9d8d-f22e149e375d" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 22 05:18:56 crc kubenswrapper[4818]: E1122 05:18:56.119825 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ab8451b-0868-4e2f-bd25-fe117c5ed373" containerName="registry-server" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.119831 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ab8451b-0868-4e2f-bd25-fe117c5ed373" containerName="registry-server" Nov 22 05:18:56 crc kubenswrapper[4818]: E1122 05:18:56.119843 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ab8451b-0868-4e2f-bd25-fe117c5ed373" containerName="extract-content" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.119849 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ab8451b-0868-4e2f-bd25-fe117c5ed373" containerName="extract-content" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.119995 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ab8451b-0868-4e2f-bd25-fe117c5ed373" containerName="registry-server" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.120015 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ffc9da3-3fd2-4752-9d8d-f22e149e375d" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.120568 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.122215 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.122289 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.122344 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.123012 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.166983 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv"] Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.210325 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d49504af-1b9b-4f48-99b1-7c5ef020f27c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-s49cv\" (UID: \"d49504af-1b9b-4f48-99b1-7c5ef020f27c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.210403 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-schcn\" (UniqueName: \"kubernetes.io/projected/d49504af-1b9b-4f48-99b1-7c5ef020f27c-kube-api-access-schcn\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-s49cv\" (UID: \"d49504af-1b9b-4f48-99b1-7c5ef020f27c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.210524 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d49504af-1b9b-4f48-99b1-7c5ef020f27c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-s49cv\" (UID: \"d49504af-1b9b-4f48-99b1-7c5ef020f27c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.311880 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-schcn\" (UniqueName: \"kubernetes.io/projected/d49504af-1b9b-4f48-99b1-7c5ef020f27c-kube-api-access-schcn\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-s49cv\" (UID: \"d49504af-1b9b-4f48-99b1-7c5ef020f27c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.312027 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d49504af-1b9b-4f48-99b1-7c5ef020f27c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-s49cv\" (UID: \"d49504af-1b9b-4f48-99b1-7c5ef020f27c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.312068 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d49504af-1b9b-4f48-99b1-7c5ef020f27c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-s49cv\" (UID: \"d49504af-1b9b-4f48-99b1-7c5ef020f27c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.317844 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d49504af-1b9b-4f48-99b1-7c5ef020f27c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-s49cv\" (UID: \"d49504af-1b9b-4f48-99b1-7c5ef020f27c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.318692 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d49504af-1b9b-4f48-99b1-7c5ef020f27c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-s49cv\" (UID: \"d49504af-1b9b-4f48-99b1-7c5ef020f27c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.329413 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-schcn\" (UniqueName: \"kubernetes.io/projected/d49504af-1b9b-4f48-99b1-7c5ef020f27c-kube-api-access-schcn\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-s49cv\" (UID: \"d49504af-1b9b-4f48-99b1-7c5ef020f27c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.446573 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv" Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.985378 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv"] Nov 22 05:18:56 crc kubenswrapper[4818]: I1122 05:18:56.989729 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 05:18:57 crc kubenswrapper[4818]: I1122 05:18:57.059758 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv" event={"ID":"d49504af-1b9b-4f48-99b1-7c5ef020f27c","Type":"ContainerStarted","Data":"ce822d24a7b588ab65955d1abb8acd27ce8a0e03b9082faafba62ff55e7153b7"} Nov 22 05:18:58 crc kubenswrapper[4818]: I1122 05:18:58.070388 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv" event={"ID":"d49504af-1b9b-4f48-99b1-7c5ef020f27c","Type":"ContainerStarted","Data":"b3c0a62a2d03a49c5a08b84ffab21d422ca50344ad57d515a03841ecfd5036a7"} Nov 22 05:19:04 crc kubenswrapper[4818]: I1122 05:19:04.291452 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:19:04 crc kubenswrapper[4818]: E1122 05:19:04.292242 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:19:19 crc kubenswrapper[4818]: I1122 05:19:19.291873 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:19:19 crc kubenswrapper[4818]: E1122 05:19:19.293440 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:19:30 crc kubenswrapper[4818]: I1122 05:19:30.292872 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:19:30 crc kubenswrapper[4818]: E1122 05:19:30.293961 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:19:37 crc kubenswrapper[4818]: I1122 05:19:37.460508 4818 scope.go:117] "RemoveContainer" containerID="291f8def5fd201497ecca5e29c50e6eed2daff4a131454883b80d38d73a92927" Nov 22 05:19:37 crc kubenswrapper[4818]: I1122 05:19:37.486720 4818 scope.go:117] "RemoveContainer" containerID="51dd0c67d382bf1b8bbe35e6cac4fe11396139cdb92c92f82e2aeab8dcf57231" Nov 22 05:19:41 crc kubenswrapper[4818]: I1122 05:19:41.292332 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:19:41 crc kubenswrapper[4818]: E1122 05:19:41.294808 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:19:53 crc kubenswrapper[4818]: I1122 05:19:53.291890 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:19:53 crc kubenswrapper[4818]: I1122 05:19:53.622982 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"d3ddef518d72589af006b70b271591f843cb53e6b99204dda0a3b0fa65c54eb9"} Nov 22 05:19:53 crc kubenswrapper[4818]: I1122 05:19:53.646802 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv" podStartSLOduration=57.226956327 podStartE2EDuration="57.646767586s" podCreationTimestamp="2025-11-22 05:18:56 +0000 UTC" firstStartedPulling="2025-11-22 05:18:56.989483651 +0000 UTC m=+1889.563900178" lastFinishedPulling="2025-11-22 05:18:57.40929491 +0000 UTC m=+1889.983711437" observedRunningTime="2025-11-22 05:18:58.091725879 +0000 UTC m=+1890.666142446" watchObservedRunningTime="2025-11-22 05:19:53.646767586 +0000 UTC m=+1946.221184153" Nov 22 05:19:54 crc kubenswrapper[4818]: I1122 05:19:54.051167 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-0c45-account-create-bcpj2"] Nov 22 05:19:54 crc kubenswrapper[4818]: I1122 05:19:54.064507 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-gmlh9"] Nov 22 05:19:54 crc kubenswrapper[4818]: I1122 05:19:54.071811 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-wg496"] Nov 22 05:19:54 crc kubenswrapper[4818]: I1122 05:19:54.078836 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-4fc8-account-create-4nkhn"] Nov 22 05:19:54 crc kubenswrapper[4818]: I1122 05:19:54.086299 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-4fc8-account-create-4nkhn"] Nov 22 05:19:54 crc kubenswrapper[4818]: I1122 05:19:54.093281 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-0c45-account-create-bcpj2"] Nov 22 05:19:54 crc kubenswrapper[4818]: I1122 05:19:54.099762 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-gmlh9"] Nov 22 05:19:54 crc kubenswrapper[4818]: I1122 05:19:54.105460 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-wg496"] Nov 22 05:19:54 crc kubenswrapper[4818]: I1122 05:19:54.311684 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a30336b-d279-4487-9748-e32b0332282b" path="/var/lib/kubelet/pods/1a30336b-d279-4487-9748-e32b0332282b/volumes" Nov 22 05:19:54 crc kubenswrapper[4818]: I1122 05:19:54.313547 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f234b24-4e7a-42dc-943d-aa45cb9cd72b" path="/var/lib/kubelet/pods/4f234b24-4e7a-42dc-943d-aa45cb9cd72b/volumes" Nov 22 05:19:54 crc kubenswrapper[4818]: I1122 05:19:54.315027 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="501b7faf-95bc-41d8-8b80-20c1c144395a" path="/var/lib/kubelet/pods/501b7faf-95bc-41d8-8b80-20c1c144395a/volumes" Nov 22 05:19:54 crc kubenswrapper[4818]: I1122 05:19:54.316577 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b0bbbc5-6ad1-4390-8d2b-b91c83f49701" path="/var/lib/kubelet/pods/5b0bbbc5-6ad1-4390-8d2b-b91c83f49701/volumes" Nov 22 05:19:55 crc kubenswrapper[4818]: I1122 05:19:55.052985 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-jbncz"] Nov 22 05:19:55 crc kubenswrapper[4818]: I1122 05:19:55.061560 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-b756-account-create-n4p6g"] Nov 22 05:19:55 crc kubenswrapper[4818]: I1122 05:19:55.068415 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-jbncz"] Nov 22 05:19:55 crc kubenswrapper[4818]: I1122 05:19:55.075396 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-b756-account-create-n4p6g"] Nov 22 05:19:56 crc kubenswrapper[4818]: I1122 05:19:56.304438 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64ae0a6e-1bb1-4db6-b021-b46253c6bbb0" path="/var/lib/kubelet/pods/64ae0a6e-1bb1-4db6-b021-b46253c6bbb0/volumes" Nov 22 05:19:56 crc kubenswrapper[4818]: I1122 05:19:56.305103 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff776fa9-dc41-4535-8a27-579d82db15b1" path="/var/lib/kubelet/pods/ff776fa9-dc41-4535-8a27-579d82db15b1/volumes" Nov 22 05:20:17 crc kubenswrapper[4818]: I1122 05:20:17.862455 4818 generic.go:334] "Generic (PLEG): container finished" podID="d49504af-1b9b-4f48-99b1-7c5ef020f27c" containerID="b3c0a62a2d03a49c5a08b84ffab21d422ca50344ad57d515a03841ecfd5036a7" exitCode=0 Nov 22 05:20:17 crc kubenswrapper[4818]: I1122 05:20:17.862539 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv" event={"ID":"d49504af-1b9b-4f48-99b1-7c5ef020f27c","Type":"ContainerDied","Data":"b3c0a62a2d03a49c5a08b84ffab21d422ca50344ad57d515a03841ecfd5036a7"} Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.284535 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv" Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.324654 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d49504af-1b9b-4f48-99b1-7c5ef020f27c-ssh-key\") pod \"d49504af-1b9b-4f48-99b1-7c5ef020f27c\" (UID: \"d49504af-1b9b-4f48-99b1-7c5ef020f27c\") " Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.324728 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-schcn\" (UniqueName: \"kubernetes.io/projected/d49504af-1b9b-4f48-99b1-7c5ef020f27c-kube-api-access-schcn\") pod \"d49504af-1b9b-4f48-99b1-7c5ef020f27c\" (UID: \"d49504af-1b9b-4f48-99b1-7c5ef020f27c\") " Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.324777 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d49504af-1b9b-4f48-99b1-7c5ef020f27c-inventory\") pod \"d49504af-1b9b-4f48-99b1-7c5ef020f27c\" (UID: \"d49504af-1b9b-4f48-99b1-7c5ef020f27c\") " Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.330588 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d49504af-1b9b-4f48-99b1-7c5ef020f27c-kube-api-access-schcn" (OuterVolumeSpecName: "kube-api-access-schcn") pod "d49504af-1b9b-4f48-99b1-7c5ef020f27c" (UID: "d49504af-1b9b-4f48-99b1-7c5ef020f27c"). InnerVolumeSpecName "kube-api-access-schcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.352850 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d49504af-1b9b-4f48-99b1-7c5ef020f27c-inventory" (OuterVolumeSpecName: "inventory") pod "d49504af-1b9b-4f48-99b1-7c5ef020f27c" (UID: "d49504af-1b9b-4f48-99b1-7c5ef020f27c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.355238 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d49504af-1b9b-4f48-99b1-7c5ef020f27c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d49504af-1b9b-4f48-99b1-7c5ef020f27c" (UID: "d49504af-1b9b-4f48-99b1-7c5ef020f27c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.434234 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d49504af-1b9b-4f48-99b1-7c5ef020f27c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.434278 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-schcn\" (UniqueName: \"kubernetes.io/projected/d49504af-1b9b-4f48-99b1-7c5ef020f27c-kube-api-access-schcn\") on node \"crc\" DevicePath \"\"" Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.434290 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d49504af-1b9b-4f48-99b1-7c5ef020f27c-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.888492 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv" event={"ID":"d49504af-1b9b-4f48-99b1-7c5ef020f27c","Type":"ContainerDied","Data":"ce822d24a7b588ab65955d1abb8acd27ce8a0e03b9082faafba62ff55e7153b7"} Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.888538 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv" Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.888562 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce822d24a7b588ab65955d1abb8acd27ce8a0e03b9082faafba62ff55e7153b7" Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.959959 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m"] Nov 22 05:20:19 crc kubenswrapper[4818]: E1122 05:20:19.965696 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d49504af-1b9b-4f48-99b1-7c5ef020f27c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.965723 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="d49504af-1b9b-4f48-99b1-7c5ef020f27c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.965927 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="d49504af-1b9b-4f48-99b1-7c5ef020f27c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.966767 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m" Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.968912 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.968950 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.970309 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.971296 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:20:19 crc kubenswrapper[4818]: I1122 05:20:19.971956 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m"] Nov 22 05:20:20 crc kubenswrapper[4818]: I1122 05:20:20.044092 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc7ffd42-a7db-44b2-851c-1a074bb23eb7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m\" (UID: \"dc7ffd42-a7db-44b2-851c-1a074bb23eb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m" Nov 22 05:20:20 crc kubenswrapper[4818]: I1122 05:20:20.044169 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nrxj\" (UniqueName: \"kubernetes.io/projected/dc7ffd42-a7db-44b2-851c-1a074bb23eb7-kube-api-access-6nrxj\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m\" (UID: \"dc7ffd42-a7db-44b2-851c-1a074bb23eb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m" Nov 22 05:20:20 crc kubenswrapper[4818]: I1122 05:20:20.044546 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc7ffd42-a7db-44b2-851c-1a074bb23eb7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m\" (UID: \"dc7ffd42-a7db-44b2-851c-1a074bb23eb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m" Nov 22 05:20:20 crc kubenswrapper[4818]: I1122 05:20:20.146063 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc7ffd42-a7db-44b2-851c-1a074bb23eb7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m\" (UID: \"dc7ffd42-a7db-44b2-851c-1a074bb23eb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m" Nov 22 05:20:20 crc kubenswrapper[4818]: I1122 05:20:20.146139 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nrxj\" (UniqueName: \"kubernetes.io/projected/dc7ffd42-a7db-44b2-851c-1a074bb23eb7-kube-api-access-6nrxj\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m\" (UID: \"dc7ffd42-a7db-44b2-851c-1a074bb23eb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m" Nov 22 05:20:20 crc kubenswrapper[4818]: I1122 05:20:20.146239 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc7ffd42-a7db-44b2-851c-1a074bb23eb7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m\" (UID: \"dc7ffd42-a7db-44b2-851c-1a074bb23eb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m" Nov 22 05:20:20 crc kubenswrapper[4818]: I1122 05:20:20.153478 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc7ffd42-a7db-44b2-851c-1a074bb23eb7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m\" (UID: \"dc7ffd42-a7db-44b2-851c-1a074bb23eb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m" Nov 22 05:20:20 crc kubenswrapper[4818]: I1122 05:20:20.155092 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc7ffd42-a7db-44b2-851c-1a074bb23eb7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m\" (UID: \"dc7ffd42-a7db-44b2-851c-1a074bb23eb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m" Nov 22 05:20:20 crc kubenswrapper[4818]: I1122 05:20:20.163440 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nrxj\" (UniqueName: \"kubernetes.io/projected/dc7ffd42-a7db-44b2-851c-1a074bb23eb7-kube-api-access-6nrxj\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m\" (UID: \"dc7ffd42-a7db-44b2-851c-1a074bb23eb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m" Nov 22 05:20:20 crc kubenswrapper[4818]: I1122 05:20:20.289304 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m" Nov 22 05:20:20 crc kubenswrapper[4818]: I1122 05:20:20.817890 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m"] Nov 22 05:20:20 crc kubenswrapper[4818]: W1122 05:20:20.827600 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc7ffd42_a7db_44b2_851c_1a074bb23eb7.slice/crio-8f09affce3fac8e226808c095075c490b5bf5d964da53104d1c03faa8ea312b9 WatchSource:0}: Error finding container 8f09affce3fac8e226808c095075c490b5bf5d964da53104d1c03faa8ea312b9: Status 404 returned error can't find the container with id 8f09affce3fac8e226808c095075c490b5bf5d964da53104d1c03faa8ea312b9 Nov 22 05:20:20 crc kubenswrapper[4818]: I1122 05:20:20.897962 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m" event={"ID":"dc7ffd42-a7db-44b2-851c-1a074bb23eb7","Type":"ContainerStarted","Data":"8f09affce3fac8e226808c095075c490b5bf5d964da53104d1c03faa8ea312b9"} Nov 22 05:20:22 crc kubenswrapper[4818]: I1122 05:20:22.935354 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m" event={"ID":"dc7ffd42-a7db-44b2-851c-1a074bb23eb7","Type":"ContainerStarted","Data":"5e903e3e975b2786aa56ea33b7ead03f01d47ccf166695a4359248206c7aada5"} Nov 22 05:20:22 crc kubenswrapper[4818]: I1122 05:20:22.960270 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m" podStartSLOduration=2.469740653 podStartE2EDuration="3.96023405s" podCreationTimestamp="2025-11-22 05:20:19 +0000 UTC" firstStartedPulling="2025-11-22 05:20:20.83004919 +0000 UTC m=+1973.404465717" lastFinishedPulling="2025-11-22 05:20:22.320542547 +0000 UTC m=+1974.894959114" observedRunningTime="2025-11-22 05:20:22.953644294 +0000 UTC m=+1975.528060851" watchObservedRunningTime="2025-11-22 05:20:22.96023405 +0000 UTC m=+1975.534650587" Nov 22 05:20:27 crc kubenswrapper[4818]: I1122 05:20:27.983438 4818 generic.go:334] "Generic (PLEG): container finished" podID="dc7ffd42-a7db-44b2-851c-1a074bb23eb7" containerID="5e903e3e975b2786aa56ea33b7ead03f01d47ccf166695a4359248206c7aada5" exitCode=0 Nov 22 05:20:27 crc kubenswrapper[4818]: I1122 05:20:27.983551 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m" event={"ID":"dc7ffd42-a7db-44b2-851c-1a074bb23eb7","Type":"ContainerDied","Data":"5e903e3e975b2786aa56ea33b7ead03f01d47ccf166695a4359248206c7aada5"} Nov 22 05:20:29 crc kubenswrapper[4818]: I1122 05:20:29.403999 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m" Nov 22 05:20:29 crc kubenswrapper[4818]: I1122 05:20:29.429922 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc7ffd42-a7db-44b2-851c-1a074bb23eb7-ssh-key\") pod \"dc7ffd42-a7db-44b2-851c-1a074bb23eb7\" (UID: \"dc7ffd42-a7db-44b2-851c-1a074bb23eb7\") " Nov 22 05:20:29 crc kubenswrapper[4818]: I1122 05:20:29.430290 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nrxj\" (UniqueName: \"kubernetes.io/projected/dc7ffd42-a7db-44b2-851c-1a074bb23eb7-kube-api-access-6nrxj\") pod \"dc7ffd42-a7db-44b2-851c-1a074bb23eb7\" (UID: \"dc7ffd42-a7db-44b2-851c-1a074bb23eb7\") " Nov 22 05:20:29 crc kubenswrapper[4818]: I1122 05:20:29.430462 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc7ffd42-a7db-44b2-851c-1a074bb23eb7-inventory\") pod \"dc7ffd42-a7db-44b2-851c-1a074bb23eb7\" (UID: \"dc7ffd42-a7db-44b2-851c-1a074bb23eb7\") " Nov 22 05:20:29 crc kubenswrapper[4818]: I1122 05:20:29.435870 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc7ffd42-a7db-44b2-851c-1a074bb23eb7-kube-api-access-6nrxj" (OuterVolumeSpecName: "kube-api-access-6nrxj") pod "dc7ffd42-a7db-44b2-851c-1a074bb23eb7" (UID: "dc7ffd42-a7db-44b2-851c-1a074bb23eb7"). InnerVolumeSpecName "kube-api-access-6nrxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:20:29 crc kubenswrapper[4818]: I1122 05:20:29.458370 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc7ffd42-a7db-44b2-851c-1a074bb23eb7-inventory" (OuterVolumeSpecName: "inventory") pod "dc7ffd42-a7db-44b2-851c-1a074bb23eb7" (UID: "dc7ffd42-a7db-44b2-851c-1a074bb23eb7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:20:29 crc kubenswrapper[4818]: I1122 05:20:29.465933 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc7ffd42-a7db-44b2-851c-1a074bb23eb7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dc7ffd42-a7db-44b2-851c-1a074bb23eb7" (UID: "dc7ffd42-a7db-44b2-851c-1a074bb23eb7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:20:29 crc kubenswrapper[4818]: I1122 05:20:29.532642 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc7ffd42-a7db-44b2-851c-1a074bb23eb7-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:20:29 crc kubenswrapper[4818]: I1122 05:20:29.532696 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nrxj\" (UniqueName: \"kubernetes.io/projected/dc7ffd42-a7db-44b2-851c-1a074bb23eb7-kube-api-access-6nrxj\") on node \"crc\" DevicePath \"\"" Nov 22 05:20:29 crc kubenswrapper[4818]: I1122 05:20:29.532712 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc7ffd42-a7db-44b2-851c-1a074bb23eb7-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.007613 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m" event={"ID":"dc7ffd42-a7db-44b2-851c-1a074bb23eb7","Type":"ContainerDied","Data":"8f09affce3fac8e226808c095075c490b5bf5d964da53104d1c03faa8ea312b9"} Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.007666 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.007668 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f09affce3fac8e226808c095075c490b5bf5d964da53104d1c03faa8ea312b9" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.068948 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk"] Nov 22 05:20:30 crc kubenswrapper[4818]: E1122 05:20:30.069316 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc7ffd42-a7db-44b2-851c-1a074bb23eb7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.069336 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc7ffd42-a7db-44b2-851c-1a074bb23eb7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.069520 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc7ffd42-a7db-44b2-851c-1a074bb23eb7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.070062 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.072171 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.072395 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.072572 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.077032 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.082128 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk"] Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.142532 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fee6592f-65ea-480e-813a-b4839af8fb95-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kczgk\" (UID: \"fee6592f-65ea-480e-813a-b4839af8fb95\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.142599 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fee6592f-65ea-480e-813a-b4839af8fb95-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kczgk\" (UID: \"fee6592f-65ea-480e-813a-b4839af8fb95\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.142869 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngzq4\" (UniqueName: \"kubernetes.io/projected/fee6592f-65ea-480e-813a-b4839af8fb95-kube-api-access-ngzq4\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kczgk\" (UID: \"fee6592f-65ea-480e-813a-b4839af8fb95\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.245119 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fee6592f-65ea-480e-813a-b4839af8fb95-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kczgk\" (UID: \"fee6592f-65ea-480e-813a-b4839af8fb95\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.245171 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fee6592f-65ea-480e-813a-b4839af8fb95-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kczgk\" (UID: \"fee6592f-65ea-480e-813a-b4839af8fb95\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.245277 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngzq4\" (UniqueName: \"kubernetes.io/projected/fee6592f-65ea-480e-813a-b4839af8fb95-kube-api-access-ngzq4\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kczgk\" (UID: \"fee6592f-65ea-480e-813a-b4839af8fb95\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.250616 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fee6592f-65ea-480e-813a-b4839af8fb95-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kczgk\" (UID: \"fee6592f-65ea-480e-813a-b4839af8fb95\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.251576 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fee6592f-65ea-480e-813a-b4839af8fb95-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kczgk\" (UID: \"fee6592f-65ea-480e-813a-b4839af8fb95\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.261855 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngzq4\" (UniqueName: \"kubernetes.io/projected/fee6592f-65ea-480e-813a-b4839af8fb95-kube-api-access-ngzq4\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kczgk\" (UID: \"fee6592f-65ea-480e-813a-b4839af8fb95\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.398696 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk" Nov 22 05:20:30 crc kubenswrapper[4818]: I1122 05:20:30.969224 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk"] Nov 22 05:20:31 crc kubenswrapper[4818]: I1122 05:20:31.018445 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk" event={"ID":"fee6592f-65ea-480e-813a-b4839af8fb95","Type":"ContainerStarted","Data":"9a1ca6d0b6fa50729e5888e1f37dd5b34db8aef93780d49be7172a318d4fa047"} Nov 22 05:20:33 crc kubenswrapper[4818]: I1122 05:20:33.043748 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk" event={"ID":"fee6592f-65ea-480e-813a-b4839af8fb95","Type":"ContainerStarted","Data":"5540a977d6599414b040c7baccf23573d7220dc64daca33947a9bfc7ab4b08aa"} Nov 22 05:20:33 crc kubenswrapper[4818]: I1122 05:20:33.071741 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk" podStartSLOduration=2.279975899 podStartE2EDuration="3.071718994s" podCreationTimestamp="2025-11-22 05:20:30 +0000 UTC" firstStartedPulling="2025-11-22 05:20:30.978718982 +0000 UTC m=+1983.553135529" lastFinishedPulling="2025-11-22 05:20:31.770462097 +0000 UTC m=+1984.344878624" observedRunningTime="2025-11-22 05:20:33.067530912 +0000 UTC m=+1985.641947489" watchObservedRunningTime="2025-11-22 05:20:33.071718994 +0000 UTC m=+1985.646135531" Nov 22 05:20:35 crc kubenswrapper[4818]: I1122 05:20:35.050981 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-d2da-account-create-zw6nl"] Nov 22 05:20:35 crc kubenswrapper[4818]: I1122 05:20:35.060195 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-ea58-account-create-rxpn5"] Nov 22 05:20:35 crc kubenswrapper[4818]: I1122 05:20:35.069386 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-4ggzd"] Nov 22 05:20:35 crc kubenswrapper[4818]: I1122 05:20:35.079071 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-zktzs"] Nov 22 05:20:35 crc kubenswrapper[4818]: I1122 05:20:35.089446 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-h7nxw"] Nov 22 05:20:35 crc kubenswrapper[4818]: I1122 05:20:35.096132 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-zktzs"] Nov 22 05:20:35 crc kubenswrapper[4818]: I1122 05:20:35.102073 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-bacb-account-create-grbqp"] Nov 22 05:20:35 crc kubenswrapper[4818]: I1122 05:20:35.108600 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-ea58-account-create-rxpn5"] Nov 22 05:20:35 crc kubenswrapper[4818]: I1122 05:20:35.114786 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-d2da-account-create-zw6nl"] Nov 22 05:20:35 crc kubenswrapper[4818]: I1122 05:20:35.121144 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-h7nxw"] Nov 22 05:20:35 crc kubenswrapper[4818]: I1122 05:20:35.136596 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-4ggzd"] Nov 22 05:20:35 crc kubenswrapper[4818]: I1122 05:20:35.145203 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-bacb-account-create-grbqp"] Nov 22 05:20:36 crc kubenswrapper[4818]: I1122 05:20:36.303129 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="147cf615-2bd3-4db4-8d5a-40711c2b8f92" path="/var/lib/kubelet/pods/147cf615-2bd3-4db4-8d5a-40711c2b8f92/volumes" Nov 22 05:20:36 crc kubenswrapper[4818]: I1122 05:20:36.303754 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60fc3ace-6591-4551-96ba-a746eacf5609" path="/var/lib/kubelet/pods/60fc3ace-6591-4551-96ba-a746eacf5609/volumes" Nov 22 05:20:36 crc kubenswrapper[4818]: I1122 05:20:36.304238 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86b3ecec-fc53-4d2d-843b-ab24ea43d82a" path="/var/lib/kubelet/pods/86b3ecec-fc53-4d2d-843b-ab24ea43d82a/volumes" Nov 22 05:20:36 crc kubenswrapper[4818]: I1122 05:20:36.304762 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9ad29a8-c449-4ac3-a30d-64979050a1b0" path="/var/lib/kubelet/pods/a9ad29a8-c449-4ac3-a30d-64979050a1b0/volumes" Nov 22 05:20:36 crc kubenswrapper[4818]: I1122 05:20:36.305793 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc03cbc0-add4-4e90-a74a-edc369ee88b9" path="/var/lib/kubelet/pods/cc03cbc0-add4-4e90-a74a-edc369ee88b9/volumes" Nov 22 05:20:36 crc kubenswrapper[4818]: I1122 05:20:36.306291 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d879cd8e-9523-47ac-acfc-75aa7fde0bc0" path="/var/lib/kubelet/pods/d879cd8e-9523-47ac-acfc-75aa7fde0bc0/volumes" Nov 22 05:20:37 crc kubenswrapper[4818]: I1122 05:20:37.557394 4818 scope.go:117] "RemoveContainer" containerID="23df6b8419f9081d2d2d35ffa2f9b3a44a1462a282262165b265e12c7befdbc8" Nov 22 05:20:37 crc kubenswrapper[4818]: I1122 05:20:37.585908 4818 scope.go:117] "RemoveContainer" containerID="36c4045266d0598dc42d139ad7a2c151990072cc88a96c5cbfcf54fa6fbae720" Nov 22 05:20:37 crc kubenswrapper[4818]: I1122 05:20:37.699846 4818 scope.go:117] "RemoveContainer" containerID="44b20ad92d8fa5547099feca6c4ddb17c309294b0a9936a67b4a6afbf04eddae" Nov 22 05:20:37 crc kubenswrapper[4818]: I1122 05:20:37.737446 4818 scope.go:117] "RemoveContainer" containerID="590bad9f87ba50586a0e48ad3beab578112c10c64bf6bd909d17882368a2103a" Nov 22 05:20:37 crc kubenswrapper[4818]: I1122 05:20:37.786242 4818 scope.go:117] "RemoveContainer" containerID="07f7237b64c16e30208f0f923aae7adde73e12131e1c38130a1593ff036fd433" Nov 22 05:20:37 crc kubenswrapper[4818]: I1122 05:20:37.824720 4818 scope.go:117] "RemoveContainer" containerID="f338a44aab1fde1e2024f62e698dea2702f40d45d9247f882f874f7f4f861e98" Nov 22 05:20:37 crc kubenswrapper[4818]: I1122 05:20:37.875731 4818 scope.go:117] "RemoveContainer" containerID="6a799794c95bebfddd1b6311071d289ac1b4c7c16d72381aa73ebacac14e3b36" Nov 22 05:20:37 crc kubenswrapper[4818]: I1122 05:20:37.894930 4818 scope.go:117] "RemoveContainer" containerID="26aec8102cb98ea8e36f0a2ee42ecc1ca4b15ea1499c6e4b0f381b676f0997da" Nov 22 05:20:37 crc kubenswrapper[4818]: I1122 05:20:37.920547 4818 scope.go:117] "RemoveContainer" containerID="e90262f97d27864977d3acc2ab0a4b913d5d787ea950db716fedd7f07af83b64" Nov 22 05:20:37 crc kubenswrapper[4818]: I1122 05:20:37.947752 4818 scope.go:117] "RemoveContainer" containerID="d703a24247a1ed221077b7214cf4452d59d409d2d95b9481af56b9187dd209b4" Nov 22 05:20:37 crc kubenswrapper[4818]: I1122 05:20:37.969759 4818 scope.go:117] "RemoveContainer" containerID="983d3e8a99dd527229e00713cbe39c9cfe0a090d0d47ecd5f7c7ee9e23d16dd3" Nov 22 05:20:38 crc kubenswrapper[4818]: I1122 05:20:38.015735 4818 scope.go:117] "RemoveContainer" containerID="294ae3c59b977c4294755965dc0a6cf3c4693284f0a330bd64b4f1cc10c48cc2" Nov 22 05:20:42 crc kubenswrapper[4818]: I1122 05:20:42.027626 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-xdpgh"] Nov 22 05:20:42 crc kubenswrapper[4818]: I1122 05:20:42.034749 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-xdpgh"] Nov 22 05:20:42 crc kubenswrapper[4818]: I1122 05:20:42.305780 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06f95278-ba08-4ccd-af26-48c0c06157e7" path="/var/lib/kubelet/pods/06f95278-ba08-4ccd-af26-48c0c06157e7/volumes" Nov 22 05:20:54 crc kubenswrapper[4818]: I1122 05:20:54.055487 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-jpkvr"] Nov 22 05:20:54 crc kubenswrapper[4818]: I1122 05:20:54.064663 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-jpkvr"] Nov 22 05:20:54 crc kubenswrapper[4818]: I1122 05:20:54.304993 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="486177c7-b0c3-45ba-a18c-79509154c9df" path="/var/lib/kubelet/pods/486177c7-b0c3-45ba-a18c-79509154c9df/volumes" Nov 22 05:21:12 crc kubenswrapper[4818]: I1122 05:21:12.626581 4818 generic.go:334] "Generic (PLEG): container finished" podID="fee6592f-65ea-480e-813a-b4839af8fb95" containerID="5540a977d6599414b040c7baccf23573d7220dc64daca33947a9bfc7ab4b08aa" exitCode=0 Nov 22 05:21:12 crc kubenswrapper[4818]: I1122 05:21:12.626712 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk" event={"ID":"fee6592f-65ea-480e-813a-b4839af8fb95","Type":"ContainerDied","Data":"5540a977d6599414b040c7baccf23573d7220dc64daca33947a9bfc7ab4b08aa"} Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.091904 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk" Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.225726 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngzq4\" (UniqueName: \"kubernetes.io/projected/fee6592f-65ea-480e-813a-b4839af8fb95-kube-api-access-ngzq4\") pod \"fee6592f-65ea-480e-813a-b4839af8fb95\" (UID: \"fee6592f-65ea-480e-813a-b4839af8fb95\") " Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.225821 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fee6592f-65ea-480e-813a-b4839af8fb95-ssh-key\") pod \"fee6592f-65ea-480e-813a-b4839af8fb95\" (UID: \"fee6592f-65ea-480e-813a-b4839af8fb95\") " Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.226061 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fee6592f-65ea-480e-813a-b4839af8fb95-inventory\") pod \"fee6592f-65ea-480e-813a-b4839af8fb95\" (UID: \"fee6592f-65ea-480e-813a-b4839af8fb95\") " Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.234539 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fee6592f-65ea-480e-813a-b4839af8fb95-kube-api-access-ngzq4" (OuterVolumeSpecName: "kube-api-access-ngzq4") pod "fee6592f-65ea-480e-813a-b4839af8fb95" (UID: "fee6592f-65ea-480e-813a-b4839af8fb95"). InnerVolumeSpecName "kube-api-access-ngzq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.250905 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fee6592f-65ea-480e-813a-b4839af8fb95-inventory" (OuterVolumeSpecName: "inventory") pod "fee6592f-65ea-480e-813a-b4839af8fb95" (UID: "fee6592f-65ea-480e-813a-b4839af8fb95"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.261709 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fee6592f-65ea-480e-813a-b4839af8fb95-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fee6592f-65ea-480e-813a-b4839af8fb95" (UID: "fee6592f-65ea-480e-813a-b4839af8fb95"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.329330 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngzq4\" (UniqueName: \"kubernetes.io/projected/fee6592f-65ea-480e-813a-b4839af8fb95-kube-api-access-ngzq4\") on node \"crc\" DevicePath \"\"" Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.329357 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fee6592f-65ea-480e-813a-b4839af8fb95-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.329376 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fee6592f-65ea-480e-813a-b4839af8fb95-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.645270 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk" event={"ID":"fee6592f-65ea-480e-813a-b4839af8fb95","Type":"ContainerDied","Data":"9a1ca6d0b6fa50729e5888e1f37dd5b34db8aef93780d49be7172a318d4fa047"} Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.645311 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a1ca6d0b6fa50729e5888e1f37dd5b34db8aef93780d49be7172a318d4fa047" Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.645328 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk" Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.747092 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p"] Nov 22 05:21:14 crc kubenswrapper[4818]: E1122 05:21:14.747841 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fee6592f-65ea-480e-813a-b4839af8fb95" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.747864 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="fee6592f-65ea-480e-813a-b4839af8fb95" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.748072 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="fee6592f-65ea-480e-813a-b4839af8fb95" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.748781 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p" Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.751382 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.752196 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.752791 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.753104 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.757947 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p"] Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.941871 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46a98f54-df60-48ae-a6ee-5affd771243e-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p\" (UID: \"46a98f54-df60-48ae-a6ee-5affd771243e\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p" Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.942130 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnw67\" (UniqueName: \"kubernetes.io/projected/46a98f54-df60-48ae-a6ee-5affd771243e-kube-api-access-vnw67\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p\" (UID: \"46a98f54-df60-48ae-a6ee-5affd771243e\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p" Nov 22 05:21:14 crc kubenswrapper[4818]: I1122 05:21:14.942382 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46a98f54-df60-48ae-a6ee-5affd771243e-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p\" (UID: \"46a98f54-df60-48ae-a6ee-5affd771243e\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p" Nov 22 05:21:15 crc kubenswrapper[4818]: I1122 05:21:15.044370 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnw67\" (UniqueName: \"kubernetes.io/projected/46a98f54-df60-48ae-a6ee-5affd771243e-kube-api-access-vnw67\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p\" (UID: \"46a98f54-df60-48ae-a6ee-5affd771243e\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p" Nov 22 05:21:15 crc kubenswrapper[4818]: I1122 05:21:15.044527 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46a98f54-df60-48ae-a6ee-5affd771243e-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p\" (UID: \"46a98f54-df60-48ae-a6ee-5affd771243e\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p" Nov 22 05:21:15 crc kubenswrapper[4818]: I1122 05:21:15.044624 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46a98f54-df60-48ae-a6ee-5affd771243e-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p\" (UID: \"46a98f54-df60-48ae-a6ee-5affd771243e\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p" Nov 22 05:21:15 crc kubenswrapper[4818]: I1122 05:21:15.049030 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46a98f54-df60-48ae-a6ee-5affd771243e-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p\" (UID: \"46a98f54-df60-48ae-a6ee-5affd771243e\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p" Nov 22 05:21:15 crc kubenswrapper[4818]: I1122 05:21:15.050695 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46a98f54-df60-48ae-a6ee-5affd771243e-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p\" (UID: \"46a98f54-df60-48ae-a6ee-5affd771243e\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p" Nov 22 05:21:15 crc kubenswrapper[4818]: I1122 05:21:15.062965 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnw67\" (UniqueName: \"kubernetes.io/projected/46a98f54-df60-48ae-a6ee-5affd771243e-kube-api-access-vnw67\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p\" (UID: \"46a98f54-df60-48ae-a6ee-5affd771243e\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p" Nov 22 05:21:15 crc kubenswrapper[4818]: I1122 05:21:15.072663 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p" Nov 22 05:21:15 crc kubenswrapper[4818]: I1122 05:21:15.403719 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p"] Nov 22 05:21:15 crc kubenswrapper[4818]: I1122 05:21:15.654975 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p" event={"ID":"46a98f54-df60-48ae-a6ee-5affd771243e","Type":"ContainerStarted","Data":"e3a066ecf614dfb5dd7207962c114940f820fe01979aad59e8f745e988806893"} Nov 22 05:21:16 crc kubenswrapper[4818]: I1122 05:21:16.664903 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p" event={"ID":"46a98f54-df60-48ae-a6ee-5affd771243e","Type":"ContainerStarted","Data":"145cc5b5499a5258289402b4d7517e67fb6c9c1cc9c7540b0d591c98cf0f1b68"} Nov 22 05:21:16 crc kubenswrapper[4818]: I1122 05:21:16.689473 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p" podStartSLOduration=2.046103855 podStartE2EDuration="2.689451926s" podCreationTimestamp="2025-11-22 05:21:14 +0000 UTC" firstStartedPulling="2025-11-22 05:21:15.409747831 +0000 UTC m=+2027.984164368" lastFinishedPulling="2025-11-22 05:21:16.053095902 +0000 UTC m=+2028.627512439" observedRunningTime="2025-11-22 05:21:16.687795222 +0000 UTC m=+2029.262211769" watchObservedRunningTime="2025-11-22 05:21:16.689451926 +0000 UTC m=+2029.263868473" Nov 22 05:21:17 crc kubenswrapper[4818]: I1122 05:21:17.031388 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-8mhxq"] Nov 22 05:21:17 crc kubenswrapper[4818]: I1122 05:21:17.040113 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-8mhxq"] Nov 22 05:21:18 crc kubenswrapper[4818]: I1122 05:21:18.027010 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-qnwqb"] Nov 22 05:21:18 crc kubenswrapper[4818]: I1122 05:21:18.035341 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-qnwqb"] Nov 22 05:21:18 crc kubenswrapper[4818]: I1122 05:21:18.302116 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad" path="/var/lib/kubelet/pods/8a3127b6-3ea9-4c8a-aa23-e1e0452d3aad/volumes" Nov 22 05:21:18 crc kubenswrapper[4818]: I1122 05:21:18.302955 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e86e2277-a801-4562-8741-84420e0d8b5e" path="/var/lib/kubelet/pods/e86e2277-a801-4562-8741-84420e0d8b5e/volumes" Nov 22 05:21:20 crc kubenswrapper[4818]: I1122 05:21:20.698050 4818 generic.go:334] "Generic (PLEG): container finished" podID="46a98f54-df60-48ae-a6ee-5affd771243e" containerID="145cc5b5499a5258289402b4d7517e67fb6c9c1cc9c7540b0d591c98cf0f1b68" exitCode=0 Nov 22 05:21:20 crc kubenswrapper[4818]: I1122 05:21:20.698167 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p" event={"ID":"46a98f54-df60-48ae-a6ee-5affd771243e","Type":"ContainerDied","Data":"145cc5b5499a5258289402b4d7517e67fb6c9c1cc9c7540b0d591c98cf0f1b68"} Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.110574 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p" Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.211296 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46a98f54-df60-48ae-a6ee-5affd771243e-ssh-key\") pod \"46a98f54-df60-48ae-a6ee-5affd771243e\" (UID: \"46a98f54-df60-48ae-a6ee-5affd771243e\") " Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.211595 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46a98f54-df60-48ae-a6ee-5affd771243e-inventory\") pod \"46a98f54-df60-48ae-a6ee-5affd771243e\" (UID: \"46a98f54-df60-48ae-a6ee-5affd771243e\") " Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.211685 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnw67\" (UniqueName: \"kubernetes.io/projected/46a98f54-df60-48ae-a6ee-5affd771243e-kube-api-access-vnw67\") pod \"46a98f54-df60-48ae-a6ee-5affd771243e\" (UID: \"46a98f54-df60-48ae-a6ee-5affd771243e\") " Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.223945 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46a98f54-df60-48ae-a6ee-5affd771243e-kube-api-access-vnw67" (OuterVolumeSpecName: "kube-api-access-vnw67") pod "46a98f54-df60-48ae-a6ee-5affd771243e" (UID: "46a98f54-df60-48ae-a6ee-5affd771243e"). InnerVolumeSpecName "kube-api-access-vnw67". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.251717 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46a98f54-df60-48ae-a6ee-5affd771243e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "46a98f54-df60-48ae-a6ee-5affd771243e" (UID: "46a98f54-df60-48ae-a6ee-5affd771243e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.259908 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46a98f54-df60-48ae-a6ee-5affd771243e-inventory" (OuterVolumeSpecName: "inventory") pod "46a98f54-df60-48ae-a6ee-5affd771243e" (UID: "46a98f54-df60-48ae-a6ee-5affd771243e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.314274 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46a98f54-df60-48ae-a6ee-5affd771243e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.314307 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46a98f54-df60-48ae-a6ee-5affd771243e-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.314317 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnw67\" (UniqueName: \"kubernetes.io/projected/46a98f54-df60-48ae-a6ee-5affd771243e-kube-api-access-vnw67\") on node \"crc\" DevicePath \"\"" Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.719751 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p" event={"ID":"46a98f54-df60-48ae-a6ee-5affd771243e","Type":"ContainerDied","Data":"e3a066ecf614dfb5dd7207962c114940f820fe01979aad59e8f745e988806893"} Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.719785 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3a066ecf614dfb5dd7207962c114940f820fe01979aad59e8f745e988806893" Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.719798 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p" Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.791755 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt"] Nov 22 05:21:22 crc kubenswrapper[4818]: E1122 05:21:22.792562 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46a98f54-df60-48ae-a6ee-5affd771243e" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.792590 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="46a98f54-df60-48ae-a6ee-5affd771243e" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.792815 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="46a98f54-df60-48ae-a6ee-5affd771243e" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.793635 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt" Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.795967 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.796076 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.796202 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.796770 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.838586 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt"] Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.925472 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f362e5c8-ad07-472c-93f2-acf126d4135a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt\" (UID: \"f362e5c8-ad07-472c-93f2-acf126d4135a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt" Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.925572 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f362e5c8-ad07-472c-93f2-acf126d4135a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt\" (UID: \"f362e5c8-ad07-472c-93f2-acf126d4135a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt" Nov 22 05:21:22 crc kubenswrapper[4818]: I1122 05:21:22.925648 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmch4\" (UniqueName: \"kubernetes.io/projected/f362e5c8-ad07-472c-93f2-acf126d4135a-kube-api-access-mmch4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt\" (UID: \"f362e5c8-ad07-472c-93f2-acf126d4135a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt" Nov 22 05:21:23 crc kubenswrapper[4818]: I1122 05:21:23.027619 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmch4\" (UniqueName: \"kubernetes.io/projected/f362e5c8-ad07-472c-93f2-acf126d4135a-kube-api-access-mmch4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt\" (UID: \"f362e5c8-ad07-472c-93f2-acf126d4135a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt" Nov 22 05:21:23 crc kubenswrapper[4818]: I1122 05:21:23.027755 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f362e5c8-ad07-472c-93f2-acf126d4135a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt\" (UID: \"f362e5c8-ad07-472c-93f2-acf126d4135a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt" Nov 22 05:21:23 crc kubenswrapper[4818]: I1122 05:21:23.027798 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f362e5c8-ad07-472c-93f2-acf126d4135a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt\" (UID: \"f362e5c8-ad07-472c-93f2-acf126d4135a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt" Nov 22 05:21:23 crc kubenswrapper[4818]: I1122 05:21:23.032231 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f362e5c8-ad07-472c-93f2-acf126d4135a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt\" (UID: \"f362e5c8-ad07-472c-93f2-acf126d4135a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt" Nov 22 05:21:23 crc kubenswrapper[4818]: I1122 05:21:23.036840 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f362e5c8-ad07-472c-93f2-acf126d4135a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt\" (UID: \"f362e5c8-ad07-472c-93f2-acf126d4135a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt" Nov 22 05:21:23 crc kubenswrapper[4818]: I1122 05:21:23.042329 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmch4\" (UniqueName: \"kubernetes.io/projected/f362e5c8-ad07-472c-93f2-acf126d4135a-kube-api-access-mmch4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt\" (UID: \"f362e5c8-ad07-472c-93f2-acf126d4135a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt" Nov 22 05:21:23 crc kubenswrapper[4818]: I1122 05:21:23.124859 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt" Nov 22 05:21:23 crc kubenswrapper[4818]: I1122 05:21:23.750783 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt"] Nov 22 05:21:24 crc kubenswrapper[4818]: I1122 05:21:24.740104 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt" event={"ID":"f362e5c8-ad07-472c-93f2-acf126d4135a","Type":"ContainerStarted","Data":"d8ae2097cf9d1f5e554d2d06c50af3b6906b2d498b8ad631c02ced6c42434b39"} Nov 22 05:21:24 crc kubenswrapper[4818]: I1122 05:21:24.740497 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt" event={"ID":"f362e5c8-ad07-472c-93f2-acf126d4135a","Type":"ContainerStarted","Data":"e281ed829b6ddd79e2991c947d2a3a4d7e39dbac2c518f64a4e10c39b40c7b62"} Nov 22 05:21:28 crc kubenswrapper[4818]: I1122 05:21:28.035011 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt" podStartSLOduration=5.608092364 podStartE2EDuration="6.034995891s" podCreationTimestamp="2025-11-22 05:21:22 +0000 UTC" firstStartedPulling="2025-11-22 05:21:23.75730635 +0000 UTC m=+2036.331722877" lastFinishedPulling="2025-11-22 05:21:24.184209857 +0000 UTC m=+2036.758626404" observedRunningTime="2025-11-22 05:21:24.764210574 +0000 UTC m=+2037.338627111" watchObservedRunningTime="2025-11-22 05:21:28.034995891 +0000 UTC m=+2040.609412418" Nov 22 05:21:28 crc kubenswrapper[4818]: I1122 05:21:28.042359 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-rd5pw"] Nov 22 05:21:28 crc kubenswrapper[4818]: I1122 05:21:28.049620 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-rd5pw"] Nov 22 05:21:28 crc kubenswrapper[4818]: I1122 05:21:28.304442 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f84522a7-f356-40c9-8ea0-13b1503993fe" path="/var/lib/kubelet/pods/f84522a7-f356-40c9-8ea0-13b1503993fe/volumes" Nov 22 05:21:33 crc kubenswrapper[4818]: I1122 05:21:33.030342 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-n2ldz"] Nov 22 05:21:33 crc kubenswrapper[4818]: I1122 05:21:33.039920 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-n2ldz"] Nov 22 05:21:34 crc kubenswrapper[4818]: I1122 05:21:34.306557 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a8a8454-074a-48c1-b1f2-f800dd6f0f0a" path="/var/lib/kubelet/pods/2a8a8454-074a-48c1-b1f2-f800dd6f0f0a/volumes" Nov 22 05:21:38 crc kubenswrapper[4818]: I1122 05:21:38.574048 4818 scope.go:117] "RemoveContainer" containerID="09d3a921e87dbfd0c0ac1db1fa0d57bdee9ada26dbcd3ab243751b335201e1e1" Nov 22 05:21:38 crc kubenswrapper[4818]: I1122 05:21:38.613121 4818 scope.go:117] "RemoveContainer" containerID="ce5d3f0ff50a531c3adbc951f081ba274f9c58f5332864a1dba917c7fa062321" Nov 22 05:21:38 crc kubenswrapper[4818]: I1122 05:21:38.656583 4818 scope.go:117] "RemoveContainer" containerID="b54e954468bed2959f1452b89b73005be137fa127ce33fe4114b82bf04b2a03b" Nov 22 05:21:38 crc kubenswrapper[4818]: I1122 05:21:38.691043 4818 scope.go:117] "RemoveContainer" containerID="566e1df2e1e8ff694af911b80125bc51fa9ebcaae0be783202453e0a0427e77f" Nov 22 05:21:38 crc kubenswrapper[4818]: I1122 05:21:38.755585 4818 scope.go:117] "RemoveContainer" containerID="7c227959c8eedfbb9e337393ec88b1c6f2cf7e6203debba75c746dae1004b069" Nov 22 05:21:38 crc kubenswrapper[4818]: I1122 05:21:38.791109 4818 scope.go:117] "RemoveContainer" containerID="52103b4c483d95017e34149bc547bcb135efd0fe9d874cd4ef26bdfe8dedd602" Nov 22 05:21:38 crc kubenswrapper[4818]: I1122 05:21:38.837330 4818 scope.go:117] "RemoveContainer" containerID="7d8c16c52256a7f940166ce1567aa8686f167a0d8de2b989b4302e9a7cb4adf9" Nov 22 05:21:38 crc kubenswrapper[4818]: I1122 05:21:38.886759 4818 scope.go:117] "RemoveContainer" containerID="a4cb577a918c0af02b86283e9fe25886815e441e60bfdf13c49010b787d99a52" Nov 22 05:21:42 crc kubenswrapper[4818]: I1122 05:21:42.038290 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-7spb7"] Nov 22 05:21:42 crc kubenswrapper[4818]: I1122 05:21:42.054868 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-7spb7"] Nov 22 05:21:42 crc kubenswrapper[4818]: I1122 05:21:42.308510 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ca1db80-7d8d-44cf-80bf-afa1c0798182" path="/var/lib/kubelet/pods/0ca1db80-7d8d-44cf-80bf-afa1c0798182/volumes" Nov 22 05:22:16 crc kubenswrapper[4818]: E1122 05:22:16.035518 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf362e5c8_ad07_472c_93f2_acf126d4135a.slice/crio-conmon-d8ae2097cf9d1f5e554d2d06c50af3b6906b2d498b8ad631c02ced6c42434b39.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf362e5c8_ad07_472c_93f2_acf126d4135a.slice/crio-d8ae2097cf9d1f5e554d2d06c50af3b6906b2d498b8ad631c02ced6c42434b39.scope\": RecentStats: unable to find data in memory cache]" Nov 22 05:22:16 crc kubenswrapper[4818]: I1122 05:22:16.290622 4818 generic.go:334] "Generic (PLEG): container finished" podID="f362e5c8-ad07-472c-93f2-acf126d4135a" containerID="d8ae2097cf9d1f5e554d2d06c50af3b6906b2d498b8ad631c02ced6c42434b39" exitCode=0 Nov 22 05:22:16 crc kubenswrapper[4818]: I1122 05:22:16.300173 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt" event={"ID":"f362e5c8-ad07-472c-93f2-acf126d4135a","Type":"ContainerDied","Data":"d8ae2097cf9d1f5e554d2d06c50af3b6906b2d498b8ad631c02ced6c42434b39"} Nov 22 05:22:17 crc kubenswrapper[4818]: I1122 05:22:17.787668 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt" Nov 22 05:22:17 crc kubenswrapper[4818]: I1122 05:22:17.941243 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmch4\" (UniqueName: \"kubernetes.io/projected/f362e5c8-ad07-472c-93f2-acf126d4135a-kube-api-access-mmch4\") pod \"f362e5c8-ad07-472c-93f2-acf126d4135a\" (UID: \"f362e5c8-ad07-472c-93f2-acf126d4135a\") " Nov 22 05:22:17 crc kubenswrapper[4818]: I1122 05:22:17.941531 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f362e5c8-ad07-472c-93f2-acf126d4135a-inventory\") pod \"f362e5c8-ad07-472c-93f2-acf126d4135a\" (UID: \"f362e5c8-ad07-472c-93f2-acf126d4135a\") " Nov 22 05:22:17 crc kubenswrapper[4818]: I1122 05:22:17.941734 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f362e5c8-ad07-472c-93f2-acf126d4135a-ssh-key\") pod \"f362e5c8-ad07-472c-93f2-acf126d4135a\" (UID: \"f362e5c8-ad07-472c-93f2-acf126d4135a\") " Nov 22 05:22:17 crc kubenswrapper[4818]: I1122 05:22:17.946790 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f362e5c8-ad07-472c-93f2-acf126d4135a-kube-api-access-mmch4" (OuterVolumeSpecName: "kube-api-access-mmch4") pod "f362e5c8-ad07-472c-93f2-acf126d4135a" (UID: "f362e5c8-ad07-472c-93f2-acf126d4135a"). InnerVolumeSpecName "kube-api-access-mmch4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:22:17 crc kubenswrapper[4818]: I1122 05:22:17.970616 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f362e5c8-ad07-472c-93f2-acf126d4135a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f362e5c8-ad07-472c-93f2-acf126d4135a" (UID: "f362e5c8-ad07-472c-93f2-acf126d4135a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:22:17 crc kubenswrapper[4818]: I1122 05:22:17.977403 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f362e5c8-ad07-472c-93f2-acf126d4135a-inventory" (OuterVolumeSpecName: "inventory") pod "f362e5c8-ad07-472c-93f2-acf126d4135a" (UID: "f362e5c8-ad07-472c-93f2-acf126d4135a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.048546 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f362e5c8-ad07-472c-93f2-acf126d4135a-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.079537 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f362e5c8-ad07-472c-93f2-acf126d4135a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.079801 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmch4\" (UniqueName: \"kubernetes.io/projected/f362e5c8-ad07-472c-93f2-acf126d4135a-kube-api-access-mmch4\") on node \"crc\" DevicePath \"\"" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.052305 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-d92ms"] Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.079984 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-8vsln"] Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.099778 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-b75c-account-create-ztmdp"] Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.106187 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cce9-account-create-qjwmx"] Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.113177 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-d92ms"] Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.119887 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-36d8-account-create-bwv78"] Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.130006 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-5njh6"] Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.136899 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-8vsln"] Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.143164 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-b75c-account-create-ztmdp"] Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.148917 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-36d8-account-create-bwv78"] Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.156035 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cce9-account-create-qjwmx"] Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.161704 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-5njh6"] Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.302978 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="057a60f1-d0da-4590-ab6c-4de6cf4c79d1" path="/var/lib/kubelet/pods/057a60f1-d0da-4590-ab6c-4de6cf4c79d1/volumes" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.303788 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="178832e0-386a-4942-9c1f-b51959d4f455" path="/var/lib/kubelet/pods/178832e0-386a-4942-9c1f-b51959d4f455/volumes" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.304518 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39310079-a73f-4fe1-854e-aacffbddcabb" path="/var/lib/kubelet/pods/39310079-a73f-4fe1-854e-aacffbddcabb/volumes" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.305290 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a04faa3-41d0-4e96-8957-097cba1b5154" path="/var/lib/kubelet/pods/5a04faa3-41d0-4e96-8957-097cba1b5154/volumes" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.306011 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e0926ba-d1e1-470f-91a0-221471d29797" path="/var/lib/kubelet/pods/6e0926ba-d1e1-470f-91a0-221471d29797/volumes" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.306766 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5382656-d791-455a-b7a4-925c2b37d399" path="/var/lib/kubelet/pods/f5382656-d791-455a-b7a4-925c2b37d399/volumes" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.322413 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt" event={"ID":"f362e5c8-ad07-472c-93f2-acf126d4135a","Type":"ContainerDied","Data":"e281ed829b6ddd79e2991c947d2a3a4d7e39dbac2c518f64a4e10c39b40c7b62"} Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.322456 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e281ed829b6ddd79e2991c947d2a3a4d7e39dbac2c518f64a4e10c39b40c7b62" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.322538 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.391652 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-b85b7"] Nov 22 05:22:18 crc kubenswrapper[4818]: E1122 05:22:18.392027 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f362e5c8-ad07-472c-93f2-acf126d4135a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.392046 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f362e5c8-ad07-472c-93f2-acf126d4135a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.392275 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f362e5c8-ad07-472c-93f2-acf126d4135a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.392967 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-b85b7" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.396042 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.396210 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.400748 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.401114 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.407370 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-b85b7"] Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.487120 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/bfcce923-6850-4c78-b11f-e71ae4c680db-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-b85b7\" (UID: \"bfcce923-6850-4c78-b11f-e71ae4c680db\") " pod="openstack/ssh-known-hosts-edpm-deployment-b85b7" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.487202 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfpnl\" (UniqueName: \"kubernetes.io/projected/bfcce923-6850-4c78-b11f-e71ae4c680db-kube-api-access-wfpnl\") pod \"ssh-known-hosts-edpm-deployment-b85b7\" (UID: \"bfcce923-6850-4c78-b11f-e71ae4c680db\") " pod="openstack/ssh-known-hosts-edpm-deployment-b85b7" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.487228 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bfcce923-6850-4c78-b11f-e71ae4c680db-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-b85b7\" (UID: \"bfcce923-6850-4c78-b11f-e71ae4c680db\") " pod="openstack/ssh-known-hosts-edpm-deployment-b85b7" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.589143 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/bfcce923-6850-4c78-b11f-e71ae4c680db-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-b85b7\" (UID: \"bfcce923-6850-4c78-b11f-e71ae4c680db\") " pod="openstack/ssh-known-hosts-edpm-deployment-b85b7" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.589313 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfpnl\" (UniqueName: \"kubernetes.io/projected/bfcce923-6850-4c78-b11f-e71ae4c680db-kube-api-access-wfpnl\") pod \"ssh-known-hosts-edpm-deployment-b85b7\" (UID: \"bfcce923-6850-4c78-b11f-e71ae4c680db\") " pod="openstack/ssh-known-hosts-edpm-deployment-b85b7" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.589376 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bfcce923-6850-4c78-b11f-e71ae4c680db-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-b85b7\" (UID: \"bfcce923-6850-4c78-b11f-e71ae4c680db\") " pod="openstack/ssh-known-hosts-edpm-deployment-b85b7" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.594192 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/bfcce923-6850-4c78-b11f-e71ae4c680db-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-b85b7\" (UID: \"bfcce923-6850-4c78-b11f-e71ae4c680db\") " pod="openstack/ssh-known-hosts-edpm-deployment-b85b7" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.594737 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bfcce923-6850-4c78-b11f-e71ae4c680db-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-b85b7\" (UID: \"bfcce923-6850-4c78-b11f-e71ae4c680db\") " pod="openstack/ssh-known-hosts-edpm-deployment-b85b7" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.607734 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfpnl\" (UniqueName: \"kubernetes.io/projected/bfcce923-6850-4c78-b11f-e71ae4c680db-kube-api-access-wfpnl\") pod \"ssh-known-hosts-edpm-deployment-b85b7\" (UID: \"bfcce923-6850-4c78-b11f-e71ae4c680db\") " pod="openstack/ssh-known-hosts-edpm-deployment-b85b7" Nov 22 05:22:18 crc kubenswrapper[4818]: I1122 05:22:18.716746 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-b85b7" Nov 22 05:22:19 crc kubenswrapper[4818]: I1122 05:22:19.790097 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-b85b7"] Nov 22 05:22:20 crc kubenswrapper[4818]: I1122 05:22:20.340129 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-b85b7" event={"ID":"bfcce923-6850-4c78-b11f-e71ae4c680db","Type":"ContainerStarted","Data":"b3896bfa374cee39a1272224922e22e72a668319318befc31c2147a5bd09e8a7"} Nov 22 05:22:21 crc kubenswrapper[4818]: I1122 05:22:21.265224 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:22:21 crc kubenswrapper[4818]: I1122 05:22:21.265672 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:22:21 crc kubenswrapper[4818]: I1122 05:22:21.353205 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-b85b7" event={"ID":"bfcce923-6850-4c78-b11f-e71ae4c680db","Type":"ContainerStarted","Data":"f69d404bd56497fa60db8a5d2d8d155570aff7bd346fa1bd1792bac0d08c2a0c"} Nov 22 05:22:21 crc kubenswrapper[4818]: I1122 05:22:21.380075 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-b85b7" podStartSLOduration=2.7496041030000002 podStartE2EDuration="3.38005647s" podCreationTimestamp="2025-11-22 05:22:18 +0000 UTC" firstStartedPulling="2025-11-22 05:22:19.798230182 +0000 UTC m=+2092.372646709" lastFinishedPulling="2025-11-22 05:22:20.428682549 +0000 UTC m=+2093.003099076" observedRunningTime="2025-11-22 05:22:21.372129868 +0000 UTC m=+2093.946546415" watchObservedRunningTime="2025-11-22 05:22:21.38005647 +0000 UTC m=+2093.954472987" Nov 22 05:22:27 crc kubenswrapper[4818]: I1122 05:22:27.411648 4818 generic.go:334] "Generic (PLEG): container finished" podID="bfcce923-6850-4c78-b11f-e71ae4c680db" containerID="f69d404bd56497fa60db8a5d2d8d155570aff7bd346fa1bd1792bac0d08c2a0c" exitCode=0 Nov 22 05:22:27 crc kubenswrapper[4818]: I1122 05:22:27.411710 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-b85b7" event={"ID":"bfcce923-6850-4c78-b11f-e71ae4c680db","Type":"ContainerDied","Data":"f69d404bd56497fa60db8a5d2d8d155570aff7bd346fa1bd1792bac0d08c2a0c"} Nov 22 05:22:28 crc kubenswrapper[4818]: I1122 05:22:28.769286 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-b85b7" Nov 22 05:22:28 crc kubenswrapper[4818]: I1122 05:22:28.822628 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bfcce923-6850-4c78-b11f-e71ae4c680db-ssh-key-openstack-edpm-ipam\") pod \"bfcce923-6850-4c78-b11f-e71ae4c680db\" (UID: \"bfcce923-6850-4c78-b11f-e71ae4c680db\") " Nov 22 05:22:28 crc kubenswrapper[4818]: I1122 05:22:28.822756 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfpnl\" (UniqueName: \"kubernetes.io/projected/bfcce923-6850-4c78-b11f-e71ae4c680db-kube-api-access-wfpnl\") pod \"bfcce923-6850-4c78-b11f-e71ae4c680db\" (UID: \"bfcce923-6850-4c78-b11f-e71ae4c680db\") " Nov 22 05:22:28 crc kubenswrapper[4818]: I1122 05:22:28.822811 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/bfcce923-6850-4c78-b11f-e71ae4c680db-inventory-0\") pod \"bfcce923-6850-4c78-b11f-e71ae4c680db\" (UID: \"bfcce923-6850-4c78-b11f-e71ae4c680db\") " Nov 22 05:22:28 crc kubenswrapper[4818]: I1122 05:22:28.828471 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfcce923-6850-4c78-b11f-e71ae4c680db-kube-api-access-wfpnl" (OuterVolumeSpecName: "kube-api-access-wfpnl") pod "bfcce923-6850-4c78-b11f-e71ae4c680db" (UID: "bfcce923-6850-4c78-b11f-e71ae4c680db"). InnerVolumeSpecName "kube-api-access-wfpnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:22:28 crc kubenswrapper[4818]: I1122 05:22:28.864503 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfcce923-6850-4c78-b11f-e71ae4c680db-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "bfcce923-6850-4c78-b11f-e71ae4c680db" (UID: "bfcce923-6850-4c78-b11f-e71ae4c680db"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:22:28 crc kubenswrapper[4818]: I1122 05:22:28.874589 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfcce923-6850-4c78-b11f-e71ae4c680db-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "bfcce923-6850-4c78-b11f-e71ae4c680db" (UID: "bfcce923-6850-4c78-b11f-e71ae4c680db"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:22:28 crc kubenswrapper[4818]: I1122 05:22:28.924962 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bfcce923-6850-4c78-b11f-e71ae4c680db-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 22 05:22:28 crc kubenswrapper[4818]: I1122 05:22:28.925001 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfpnl\" (UniqueName: \"kubernetes.io/projected/bfcce923-6850-4c78-b11f-e71ae4c680db-kube-api-access-wfpnl\") on node \"crc\" DevicePath \"\"" Nov 22 05:22:28 crc kubenswrapper[4818]: I1122 05:22:28.925011 4818 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/bfcce923-6850-4c78-b11f-e71ae4c680db-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.431535 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-b85b7" event={"ID":"bfcce923-6850-4c78-b11f-e71ae4c680db","Type":"ContainerDied","Data":"b3896bfa374cee39a1272224922e22e72a668319318befc31c2147a5bd09e8a7"} Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.431588 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3896bfa374cee39a1272224922e22e72a668319318befc31c2147a5bd09e8a7" Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.431607 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-b85b7" Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.537496 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h"] Nov 22 05:22:29 crc kubenswrapper[4818]: E1122 05:22:29.538021 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfcce923-6850-4c78-b11f-e71ae4c680db" containerName="ssh-known-hosts-edpm-deployment" Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.538052 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfcce923-6850-4c78-b11f-e71ae4c680db" containerName="ssh-known-hosts-edpm-deployment" Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.539390 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfcce923-6850-4c78-b11f-e71ae4c680db" containerName="ssh-known-hosts-edpm-deployment" Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.540187 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h" Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.541886 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.543565 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.543958 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.549234 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.551287 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h"] Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.642961 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79833bd6-1636-441a-8fae-e319a7d0873e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ml85h\" (UID: \"79833bd6-1636-441a-8fae-e319a7d0873e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h" Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.643036 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79833bd6-1636-441a-8fae-e319a7d0873e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ml85h\" (UID: \"79833bd6-1636-441a-8fae-e319a7d0873e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h" Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.643178 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs69h\" (UniqueName: \"kubernetes.io/projected/79833bd6-1636-441a-8fae-e319a7d0873e-kube-api-access-bs69h\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ml85h\" (UID: \"79833bd6-1636-441a-8fae-e319a7d0873e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h" Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.745285 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79833bd6-1636-441a-8fae-e319a7d0873e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ml85h\" (UID: \"79833bd6-1636-441a-8fae-e319a7d0873e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h" Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.745331 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79833bd6-1636-441a-8fae-e319a7d0873e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ml85h\" (UID: \"79833bd6-1636-441a-8fae-e319a7d0873e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h" Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.745377 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs69h\" (UniqueName: \"kubernetes.io/projected/79833bd6-1636-441a-8fae-e319a7d0873e-kube-api-access-bs69h\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ml85h\" (UID: \"79833bd6-1636-441a-8fae-e319a7d0873e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h" Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.749085 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79833bd6-1636-441a-8fae-e319a7d0873e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ml85h\" (UID: \"79833bd6-1636-441a-8fae-e319a7d0873e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h" Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.749133 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79833bd6-1636-441a-8fae-e319a7d0873e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ml85h\" (UID: \"79833bd6-1636-441a-8fae-e319a7d0873e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h" Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.774865 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs69h\" (UniqueName: \"kubernetes.io/projected/79833bd6-1636-441a-8fae-e319a7d0873e-kube-api-access-bs69h\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ml85h\" (UID: \"79833bd6-1636-441a-8fae-e319a7d0873e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h" Nov 22 05:22:29 crc kubenswrapper[4818]: I1122 05:22:29.857972 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h" Nov 22 05:22:30 crc kubenswrapper[4818]: I1122 05:22:30.440138 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h"] Nov 22 05:22:31 crc kubenswrapper[4818]: I1122 05:22:31.454620 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h" event={"ID":"79833bd6-1636-441a-8fae-e319a7d0873e","Type":"ContainerStarted","Data":"c95b0ec7b59f1472d7b7b8f88d063226e609dd63781d5ab875d78d7c5197480e"} Nov 22 05:22:31 crc kubenswrapper[4818]: I1122 05:22:31.455660 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h" event={"ID":"79833bd6-1636-441a-8fae-e319a7d0873e","Type":"ContainerStarted","Data":"ee7e15c73df984f9ad8bf7a6dcc26c974b3676b9659db72a1a0b633692bb3b84"} Nov 22 05:22:31 crc kubenswrapper[4818]: I1122 05:22:31.479828 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h" podStartSLOduration=1.9585231300000001 podStartE2EDuration="2.479801626s" podCreationTimestamp="2025-11-22 05:22:29 +0000 UTC" firstStartedPulling="2025-11-22 05:22:30.449669428 +0000 UTC m=+2103.024085955" lastFinishedPulling="2025-11-22 05:22:30.970947894 +0000 UTC m=+2103.545364451" observedRunningTime="2025-11-22 05:22:31.475859242 +0000 UTC m=+2104.050275799" watchObservedRunningTime="2025-11-22 05:22:31.479801626 +0000 UTC m=+2104.054218153" Nov 22 05:22:39 crc kubenswrapper[4818]: I1122 05:22:39.027557 4818 scope.go:117] "RemoveContainer" containerID="d060af874680974cd4ce7851cfd0125d9efce86e4558652834b2cec0ba64c7fa" Nov 22 05:22:39 crc kubenswrapper[4818]: I1122 05:22:39.051487 4818 scope.go:117] "RemoveContainer" containerID="d4c091ded479b158d589b6c9415e9c32744493ea4418f8cc86e5a1d57e0fcb33" Nov 22 05:22:39 crc kubenswrapper[4818]: I1122 05:22:39.098655 4818 scope.go:117] "RemoveContainer" containerID="320c6f3306f348ea0711da8ae7f46f05fead8dbe60d4fc73c16afe52ead37202" Nov 22 05:22:39 crc kubenswrapper[4818]: I1122 05:22:39.159134 4818 scope.go:117] "RemoveContainer" containerID="f91a1b449e7bb6c68f7e492764683b0d1a2ab0b6e3abcd72b853dd558b3eeb8c" Nov 22 05:22:39 crc kubenswrapper[4818]: I1122 05:22:39.195935 4818 scope.go:117] "RemoveContainer" containerID="3f03b336d1afc3c9ecdd9f5afb3b508da56c440e205176a76eeb5c1be9576b09" Nov 22 05:22:39 crc kubenswrapper[4818]: I1122 05:22:39.233977 4818 scope.go:117] "RemoveContainer" containerID="ded3d607ac66ac24edd63abaae5ebd0d99640499448085fa891e07777e9b7753" Nov 22 05:22:39 crc kubenswrapper[4818]: I1122 05:22:39.278194 4818 scope.go:117] "RemoveContainer" containerID="fed7bf830da4016586061964fd8e2a8b04bd20aa622442114222ec6e1297420a" Nov 22 05:22:40 crc kubenswrapper[4818]: I1122 05:22:40.549591 4818 generic.go:334] "Generic (PLEG): container finished" podID="79833bd6-1636-441a-8fae-e319a7d0873e" containerID="c95b0ec7b59f1472d7b7b8f88d063226e609dd63781d5ab875d78d7c5197480e" exitCode=0 Nov 22 05:22:40 crc kubenswrapper[4818]: I1122 05:22:40.549654 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h" event={"ID":"79833bd6-1636-441a-8fae-e319a7d0873e","Type":"ContainerDied","Data":"c95b0ec7b59f1472d7b7b8f88d063226e609dd63781d5ab875d78d7c5197480e"} Nov 22 05:22:41 crc kubenswrapper[4818]: I1122 05:22:41.939393 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h" Nov 22 05:22:41 crc kubenswrapper[4818]: I1122 05:22:41.986432 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bs69h\" (UniqueName: \"kubernetes.io/projected/79833bd6-1636-441a-8fae-e319a7d0873e-kube-api-access-bs69h\") pod \"79833bd6-1636-441a-8fae-e319a7d0873e\" (UID: \"79833bd6-1636-441a-8fae-e319a7d0873e\") " Nov 22 05:22:41 crc kubenswrapper[4818]: I1122 05:22:41.986509 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79833bd6-1636-441a-8fae-e319a7d0873e-inventory\") pod \"79833bd6-1636-441a-8fae-e319a7d0873e\" (UID: \"79833bd6-1636-441a-8fae-e319a7d0873e\") " Nov 22 05:22:41 crc kubenswrapper[4818]: I1122 05:22:41.986545 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79833bd6-1636-441a-8fae-e319a7d0873e-ssh-key\") pod \"79833bd6-1636-441a-8fae-e319a7d0873e\" (UID: \"79833bd6-1636-441a-8fae-e319a7d0873e\") " Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.012182 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79833bd6-1636-441a-8fae-e319a7d0873e-kube-api-access-bs69h" (OuterVolumeSpecName: "kube-api-access-bs69h") pod "79833bd6-1636-441a-8fae-e319a7d0873e" (UID: "79833bd6-1636-441a-8fae-e319a7d0873e"). InnerVolumeSpecName "kube-api-access-bs69h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.016437 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79833bd6-1636-441a-8fae-e319a7d0873e-inventory" (OuterVolumeSpecName: "inventory") pod "79833bd6-1636-441a-8fae-e319a7d0873e" (UID: "79833bd6-1636-441a-8fae-e319a7d0873e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.022104 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79833bd6-1636-441a-8fae-e319a7d0873e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "79833bd6-1636-441a-8fae-e319a7d0873e" (UID: "79833bd6-1636-441a-8fae-e319a7d0873e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.088596 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bs69h\" (UniqueName: \"kubernetes.io/projected/79833bd6-1636-441a-8fae-e319a7d0873e-kube-api-access-bs69h\") on node \"crc\" DevicePath \"\"" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.088630 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79833bd6-1636-441a-8fae-e319a7d0873e-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.088639 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79833bd6-1636-441a-8fae-e319a7d0873e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.602489 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h" event={"ID":"79833bd6-1636-441a-8fae-e319a7d0873e","Type":"ContainerDied","Data":"ee7e15c73df984f9ad8bf7a6dcc26c974b3676b9659db72a1a0b633692bb3b84"} Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.602540 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee7e15c73df984f9ad8bf7a6dcc26c974b3676b9659db72a1a0b633692bb3b84" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.602595 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.666189 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc"] Nov 22 05:22:42 crc kubenswrapper[4818]: E1122 05:22:42.666660 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79833bd6-1636-441a-8fae-e319a7d0873e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.666684 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="79833bd6-1636-441a-8fae-e319a7d0873e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.666928 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="79833bd6-1636-441a-8fae-e319a7d0873e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.667795 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.671704 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.675594 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.675777 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.676274 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.678940 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc"] Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.699860 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ce7c513-156b-4380-a2da-1a2ab6ee32e4-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc\" (UID: \"7ce7c513-156b-4380-a2da-1a2ab6ee32e4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.699967 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ce7c513-156b-4380-a2da-1a2ab6ee32e4-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc\" (UID: \"7ce7c513-156b-4380-a2da-1a2ab6ee32e4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.700090 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjf5j\" (UniqueName: \"kubernetes.io/projected/7ce7c513-156b-4380-a2da-1a2ab6ee32e4-kube-api-access-fjf5j\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc\" (UID: \"7ce7c513-156b-4380-a2da-1a2ab6ee32e4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.801154 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjf5j\" (UniqueName: \"kubernetes.io/projected/7ce7c513-156b-4380-a2da-1a2ab6ee32e4-kube-api-access-fjf5j\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc\" (UID: \"7ce7c513-156b-4380-a2da-1a2ab6ee32e4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.801360 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ce7c513-156b-4380-a2da-1a2ab6ee32e4-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc\" (UID: \"7ce7c513-156b-4380-a2da-1a2ab6ee32e4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.801402 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ce7c513-156b-4380-a2da-1a2ab6ee32e4-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc\" (UID: \"7ce7c513-156b-4380-a2da-1a2ab6ee32e4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.805604 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ce7c513-156b-4380-a2da-1a2ab6ee32e4-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc\" (UID: \"7ce7c513-156b-4380-a2da-1a2ab6ee32e4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.807001 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ce7c513-156b-4380-a2da-1a2ab6ee32e4-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc\" (UID: \"7ce7c513-156b-4380-a2da-1a2ab6ee32e4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.822945 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjf5j\" (UniqueName: \"kubernetes.io/projected/7ce7c513-156b-4380-a2da-1a2ab6ee32e4-kube-api-access-fjf5j\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc\" (UID: \"7ce7c513-156b-4380-a2da-1a2ab6ee32e4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc" Nov 22 05:22:42 crc kubenswrapper[4818]: I1122 05:22:42.992484 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc" Nov 22 05:22:43 crc kubenswrapper[4818]: I1122 05:22:43.600611 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc"] Nov 22 05:22:44 crc kubenswrapper[4818]: I1122 05:22:44.622439 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc" event={"ID":"7ce7c513-156b-4380-a2da-1a2ab6ee32e4","Type":"ContainerStarted","Data":"afe7820a66f248564a7448379bba8a9560929ca8857b3a4606b77acaa5938d27"} Nov 22 05:22:44 crc kubenswrapper[4818]: I1122 05:22:44.622905 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc" event={"ID":"7ce7c513-156b-4380-a2da-1a2ab6ee32e4","Type":"ContainerStarted","Data":"fad319347dc88f4af038a4a5a9935da8cac11dcbda49d6292bae1954e8f8a956"} Nov 22 05:22:44 crc kubenswrapper[4818]: I1122 05:22:44.641236 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc" podStartSLOduration=2.198989249 podStartE2EDuration="2.641217079s" podCreationTimestamp="2025-11-22 05:22:42 +0000 UTC" firstStartedPulling="2025-11-22 05:22:43.613272679 +0000 UTC m=+2116.187689216" lastFinishedPulling="2025-11-22 05:22:44.055500519 +0000 UTC m=+2116.629917046" observedRunningTime="2025-11-22 05:22:44.635199818 +0000 UTC m=+2117.209616355" watchObservedRunningTime="2025-11-22 05:22:44.641217079 +0000 UTC m=+2117.215633606" Nov 22 05:22:45 crc kubenswrapper[4818]: I1122 05:22:45.042343 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xwlll"] Nov 22 05:22:45 crc kubenswrapper[4818]: I1122 05:22:45.050955 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xwlll"] Nov 22 05:22:46 crc kubenswrapper[4818]: I1122 05:22:46.302766 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00fbaf9a-fd45-4f17-a114-b64695d2456d" path="/var/lib/kubelet/pods/00fbaf9a-fd45-4f17-a114-b64695d2456d/volumes" Nov 22 05:22:51 crc kubenswrapper[4818]: I1122 05:22:51.265315 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:22:51 crc kubenswrapper[4818]: I1122 05:22:51.265917 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:22:54 crc kubenswrapper[4818]: I1122 05:22:54.713137 4818 generic.go:334] "Generic (PLEG): container finished" podID="7ce7c513-156b-4380-a2da-1a2ab6ee32e4" containerID="afe7820a66f248564a7448379bba8a9560929ca8857b3a4606b77acaa5938d27" exitCode=0 Nov 22 05:22:54 crc kubenswrapper[4818]: I1122 05:22:54.713205 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc" event={"ID":"7ce7c513-156b-4380-a2da-1a2ab6ee32e4","Type":"ContainerDied","Data":"afe7820a66f248564a7448379bba8a9560929ca8857b3a4606b77acaa5938d27"} Nov 22 05:22:56 crc kubenswrapper[4818]: I1122 05:22:56.193621 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc" Nov 22 05:22:56 crc kubenswrapper[4818]: I1122 05:22:56.308473 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjf5j\" (UniqueName: \"kubernetes.io/projected/7ce7c513-156b-4380-a2da-1a2ab6ee32e4-kube-api-access-fjf5j\") pod \"7ce7c513-156b-4380-a2da-1a2ab6ee32e4\" (UID: \"7ce7c513-156b-4380-a2da-1a2ab6ee32e4\") " Nov 22 05:22:56 crc kubenswrapper[4818]: I1122 05:22:56.308668 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ce7c513-156b-4380-a2da-1a2ab6ee32e4-ssh-key\") pod \"7ce7c513-156b-4380-a2da-1a2ab6ee32e4\" (UID: \"7ce7c513-156b-4380-a2da-1a2ab6ee32e4\") " Nov 22 05:22:56 crc kubenswrapper[4818]: I1122 05:22:56.308827 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ce7c513-156b-4380-a2da-1a2ab6ee32e4-inventory\") pod \"7ce7c513-156b-4380-a2da-1a2ab6ee32e4\" (UID: \"7ce7c513-156b-4380-a2da-1a2ab6ee32e4\") " Nov 22 05:22:56 crc kubenswrapper[4818]: I1122 05:22:56.318596 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ce7c513-156b-4380-a2da-1a2ab6ee32e4-kube-api-access-fjf5j" (OuterVolumeSpecName: "kube-api-access-fjf5j") pod "7ce7c513-156b-4380-a2da-1a2ab6ee32e4" (UID: "7ce7c513-156b-4380-a2da-1a2ab6ee32e4"). InnerVolumeSpecName "kube-api-access-fjf5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:22:56 crc kubenswrapper[4818]: I1122 05:22:56.337991 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ce7c513-156b-4380-a2da-1a2ab6ee32e4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7ce7c513-156b-4380-a2da-1a2ab6ee32e4" (UID: "7ce7c513-156b-4380-a2da-1a2ab6ee32e4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:22:56 crc kubenswrapper[4818]: I1122 05:22:56.355676 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ce7c513-156b-4380-a2da-1a2ab6ee32e4-inventory" (OuterVolumeSpecName: "inventory") pod "7ce7c513-156b-4380-a2da-1a2ab6ee32e4" (UID: "7ce7c513-156b-4380-a2da-1a2ab6ee32e4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:22:56 crc kubenswrapper[4818]: I1122 05:22:56.412000 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjf5j\" (UniqueName: \"kubernetes.io/projected/7ce7c513-156b-4380-a2da-1a2ab6ee32e4-kube-api-access-fjf5j\") on node \"crc\" DevicePath \"\"" Nov 22 05:22:56 crc kubenswrapper[4818]: I1122 05:22:56.412077 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ce7c513-156b-4380-a2da-1a2ab6ee32e4-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:22:56 crc kubenswrapper[4818]: I1122 05:22:56.412103 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ce7c513-156b-4380-a2da-1a2ab6ee32e4-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:22:56 crc kubenswrapper[4818]: I1122 05:22:56.736736 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc" event={"ID":"7ce7c513-156b-4380-a2da-1a2ab6ee32e4","Type":"ContainerDied","Data":"fad319347dc88f4af038a4a5a9935da8cac11dcbda49d6292bae1954e8f8a956"} Nov 22 05:22:56 crc kubenswrapper[4818]: I1122 05:22:56.737018 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fad319347dc88f4af038a4a5a9935da8cac11dcbda49d6292bae1954e8f8a956" Nov 22 05:22:56 crc kubenswrapper[4818]: I1122 05:22:56.736771 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc" Nov 22 05:23:07 crc kubenswrapper[4818]: I1122 05:23:07.052205 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-f9mzx"] Nov 22 05:23:07 crc kubenswrapper[4818]: I1122 05:23:07.061103 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-f9mzx"] Nov 22 05:23:08 crc kubenswrapper[4818]: I1122 05:23:08.303864 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7a2b652-b72b-4cd4-85cb-37e996f5238c" path="/var/lib/kubelet/pods/c7a2b652-b72b-4cd4-85cb-37e996f5238c/volumes" Nov 22 05:23:11 crc kubenswrapper[4818]: I1122 05:23:11.038453 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tfdvb"] Nov 22 05:23:11 crc kubenswrapper[4818]: I1122 05:23:11.052324 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tfdvb"] Nov 22 05:23:12 crc kubenswrapper[4818]: I1122 05:23:12.326628 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10dfb213-2885-405b-8c78-78363efec183" path="/var/lib/kubelet/pods/10dfb213-2885-405b-8c78-78363efec183/volumes" Nov 22 05:23:17 crc kubenswrapper[4818]: I1122 05:23:17.185464 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lz4zd"] Nov 22 05:23:17 crc kubenswrapper[4818]: E1122 05:23:17.186412 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce7c513-156b-4380-a2da-1a2ab6ee32e4" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:23:17 crc kubenswrapper[4818]: I1122 05:23:17.186431 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce7c513-156b-4380-a2da-1a2ab6ee32e4" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:23:17 crc kubenswrapper[4818]: I1122 05:23:17.186628 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ce7c513-156b-4380-a2da-1a2ab6ee32e4" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:23:17 crc kubenswrapper[4818]: I1122 05:23:17.188219 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lz4zd" Nov 22 05:23:17 crc kubenswrapper[4818]: I1122 05:23:17.200860 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lz4zd"] Nov 22 05:23:17 crc kubenswrapper[4818]: I1122 05:23:17.352377 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l27v9\" (UniqueName: \"kubernetes.io/projected/793ce69e-3676-4cdd-8b9a-97164a8399ff-kube-api-access-l27v9\") pod \"redhat-operators-lz4zd\" (UID: \"793ce69e-3676-4cdd-8b9a-97164a8399ff\") " pod="openshift-marketplace/redhat-operators-lz4zd" Nov 22 05:23:17 crc kubenswrapper[4818]: I1122 05:23:17.352818 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/793ce69e-3676-4cdd-8b9a-97164a8399ff-utilities\") pod \"redhat-operators-lz4zd\" (UID: \"793ce69e-3676-4cdd-8b9a-97164a8399ff\") " pod="openshift-marketplace/redhat-operators-lz4zd" Nov 22 05:23:17 crc kubenswrapper[4818]: I1122 05:23:17.352866 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/793ce69e-3676-4cdd-8b9a-97164a8399ff-catalog-content\") pod \"redhat-operators-lz4zd\" (UID: \"793ce69e-3676-4cdd-8b9a-97164a8399ff\") " pod="openshift-marketplace/redhat-operators-lz4zd" Nov 22 05:23:17 crc kubenswrapper[4818]: I1122 05:23:17.453710 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l27v9\" (UniqueName: \"kubernetes.io/projected/793ce69e-3676-4cdd-8b9a-97164a8399ff-kube-api-access-l27v9\") pod \"redhat-operators-lz4zd\" (UID: \"793ce69e-3676-4cdd-8b9a-97164a8399ff\") " pod="openshift-marketplace/redhat-operators-lz4zd" Nov 22 05:23:17 crc kubenswrapper[4818]: I1122 05:23:17.454799 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/793ce69e-3676-4cdd-8b9a-97164a8399ff-utilities\") pod \"redhat-operators-lz4zd\" (UID: \"793ce69e-3676-4cdd-8b9a-97164a8399ff\") " pod="openshift-marketplace/redhat-operators-lz4zd" Nov 22 05:23:17 crc kubenswrapper[4818]: I1122 05:23:17.455094 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/793ce69e-3676-4cdd-8b9a-97164a8399ff-catalog-content\") pod \"redhat-operators-lz4zd\" (UID: \"793ce69e-3676-4cdd-8b9a-97164a8399ff\") " pod="openshift-marketplace/redhat-operators-lz4zd" Nov 22 05:23:17 crc kubenswrapper[4818]: I1122 05:23:17.455422 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/793ce69e-3676-4cdd-8b9a-97164a8399ff-utilities\") pod \"redhat-operators-lz4zd\" (UID: \"793ce69e-3676-4cdd-8b9a-97164a8399ff\") " pod="openshift-marketplace/redhat-operators-lz4zd" Nov 22 05:23:17 crc kubenswrapper[4818]: I1122 05:23:17.455474 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/793ce69e-3676-4cdd-8b9a-97164a8399ff-catalog-content\") pod \"redhat-operators-lz4zd\" (UID: \"793ce69e-3676-4cdd-8b9a-97164a8399ff\") " pod="openshift-marketplace/redhat-operators-lz4zd" Nov 22 05:23:17 crc kubenswrapper[4818]: I1122 05:23:17.474589 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l27v9\" (UniqueName: \"kubernetes.io/projected/793ce69e-3676-4cdd-8b9a-97164a8399ff-kube-api-access-l27v9\") pod \"redhat-operators-lz4zd\" (UID: \"793ce69e-3676-4cdd-8b9a-97164a8399ff\") " pod="openshift-marketplace/redhat-operators-lz4zd" Nov 22 05:23:17 crc kubenswrapper[4818]: I1122 05:23:17.514813 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lz4zd" Nov 22 05:23:18 crc kubenswrapper[4818]: I1122 05:23:18.010394 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lz4zd"] Nov 22 05:23:18 crc kubenswrapper[4818]: I1122 05:23:18.938161 4818 generic.go:334] "Generic (PLEG): container finished" podID="793ce69e-3676-4cdd-8b9a-97164a8399ff" containerID="c83d54a48469030e65a2c53de994ca64e8bb882ad5e9dc4142e462eff87e491d" exitCode=0 Nov 22 05:23:18 crc kubenswrapper[4818]: I1122 05:23:18.938217 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lz4zd" event={"ID":"793ce69e-3676-4cdd-8b9a-97164a8399ff","Type":"ContainerDied","Data":"c83d54a48469030e65a2c53de994ca64e8bb882ad5e9dc4142e462eff87e491d"} Nov 22 05:23:18 crc kubenswrapper[4818]: I1122 05:23:18.939440 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lz4zd" event={"ID":"793ce69e-3676-4cdd-8b9a-97164a8399ff","Type":"ContainerStarted","Data":"ef7a88f8846a9640e960ab9d6a90e931fd0df46b96fcb49cb8c8d59755aeca85"} Nov 22 05:23:20 crc kubenswrapper[4818]: I1122 05:23:20.966230 4818 generic.go:334] "Generic (PLEG): container finished" podID="793ce69e-3676-4cdd-8b9a-97164a8399ff" containerID="64f66b41feda13b421a00a0947b33d8dbf8f07b1de0d50c0f0be7d21f641c7fe" exitCode=0 Nov 22 05:23:20 crc kubenswrapper[4818]: I1122 05:23:20.966740 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lz4zd" event={"ID":"793ce69e-3676-4cdd-8b9a-97164a8399ff","Type":"ContainerDied","Data":"64f66b41feda13b421a00a0947b33d8dbf8f07b1de0d50c0f0be7d21f641c7fe"} Nov 22 05:23:21 crc kubenswrapper[4818]: I1122 05:23:21.265310 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:23:21 crc kubenswrapper[4818]: I1122 05:23:21.265675 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:23:21 crc kubenswrapper[4818]: I1122 05:23:21.265739 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 05:23:21 crc kubenswrapper[4818]: I1122 05:23:21.266679 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d3ddef518d72589af006b70b271591f843cb53e6b99204dda0a3b0fa65c54eb9"} pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 05:23:21 crc kubenswrapper[4818]: I1122 05:23:21.266774 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" containerID="cri-o://d3ddef518d72589af006b70b271591f843cb53e6b99204dda0a3b0fa65c54eb9" gracePeriod=600 Nov 22 05:23:21 crc kubenswrapper[4818]: I1122 05:23:21.976127 4818 generic.go:334] "Generic (PLEG): container finished" podID="0971c9ec-29ce-44eb-9875-ad4664143305" containerID="d3ddef518d72589af006b70b271591f843cb53e6b99204dda0a3b0fa65c54eb9" exitCode=0 Nov 22 05:23:21 crc kubenswrapper[4818]: I1122 05:23:21.976181 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerDied","Data":"d3ddef518d72589af006b70b271591f843cb53e6b99204dda0a3b0fa65c54eb9"} Nov 22 05:23:21 crc kubenswrapper[4818]: I1122 05:23:21.976227 4818 scope.go:117] "RemoveContainer" containerID="9b687e2dc9a77afeecb454fae8e0acd4c3189d931b7d095853a43286d188f522" Nov 22 05:23:22 crc kubenswrapper[4818]: I1122 05:23:22.985725 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lz4zd" event={"ID":"793ce69e-3676-4cdd-8b9a-97164a8399ff","Type":"ContainerStarted","Data":"80f6c106ab3fd9dbe8622602a0b326bc043a1727e585723b94f2f43c3078fd78"} Nov 22 05:23:23 crc kubenswrapper[4818]: I1122 05:23:23.002782 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc"} Nov 22 05:23:23 crc kubenswrapper[4818]: I1122 05:23:23.039963 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lz4zd" podStartSLOduration=2.762952829 podStartE2EDuration="6.039933656s" podCreationTimestamp="2025-11-22 05:23:17 +0000 UTC" firstStartedPulling="2025-11-22 05:23:18.941241386 +0000 UTC m=+2151.515657943" lastFinishedPulling="2025-11-22 05:23:22.218222203 +0000 UTC m=+2154.792638770" observedRunningTime="2025-11-22 05:23:23.021165095 +0000 UTC m=+2155.595581632" watchObservedRunningTime="2025-11-22 05:23:23.039933656 +0000 UTC m=+2155.614350213" Nov 22 05:23:27 crc kubenswrapper[4818]: I1122 05:23:27.515059 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lz4zd" Nov 22 05:23:27 crc kubenswrapper[4818]: I1122 05:23:27.515462 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lz4zd" Nov 22 05:23:28 crc kubenswrapper[4818]: I1122 05:23:28.581784 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lz4zd" podUID="793ce69e-3676-4cdd-8b9a-97164a8399ff" containerName="registry-server" probeResult="failure" output=< Nov 22 05:23:28 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Nov 22 05:23:28 crc kubenswrapper[4818]: > Nov 22 05:23:37 crc kubenswrapper[4818]: I1122 05:23:37.578284 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lz4zd" Nov 22 05:23:37 crc kubenswrapper[4818]: I1122 05:23:37.642636 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lz4zd" Nov 22 05:23:38 crc kubenswrapper[4818]: I1122 05:23:38.054889 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lz4zd"] Nov 22 05:23:39 crc kubenswrapper[4818]: I1122 05:23:39.166666 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lz4zd" podUID="793ce69e-3676-4cdd-8b9a-97164a8399ff" containerName="registry-server" containerID="cri-o://80f6c106ab3fd9dbe8622602a0b326bc043a1727e585723b94f2f43c3078fd78" gracePeriod=2 Nov 22 05:23:39 crc kubenswrapper[4818]: I1122 05:23:39.422660 4818 scope.go:117] "RemoveContainer" containerID="4f8a0beee11a41b125ccd8bcedeca4bb9b6a58dc64b4e10d3c2f83b5853f9851" Nov 22 05:23:39 crc kubenswrapper[4818]: I1122 05:23:39.477728 4818 scope.go:117] "RemoveContainer" containerID="3a41077bbf785cb7d8ba47f08356c586c5e42d79788b494a74d45edd2c6ea260" Nov 22 05:23:39 crc kubenswrapper[4818]: I1122 05:23:39.524048 4818 scope.go:117] "RemoveContainer" containerID="cabefd1cca18ebbda0e78b29b6efb1274346a30db9cdbd8a9d7dca2b2bb35068" Nov 22 05:23:39 crc kubenswrapper[4818]: I1122 05:23:39.615307 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lz4zd" Nov 22 05:23:39 crc kubenswrapper[4818]: I1122 05:23:39.706861 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l27v9\" (UniqueName: \"kubernetes.io/projected/793ce69e-3676-4cdd-8b9a-97164a8399ff-kube-api-access-l27v9\") pod \"793ce69e-3676-4cdd-8b9a-97164a8399ff\" (UID: \"793ce69e-3676-4cdd-8b9a-97164a8399ff\") " Nov 22 05:23:39 crc kubenswrapper[4818]: I1122 05:23:39.707061 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/793ce69e-3676-4cdd-8b9a-97164a8399ff-catalog-content\") pod \"793ce69e-3676-4cdd-8b9a-97164a8399ff\" (UID: \"793ce69e-3676-4cdd-8b9a-97164a8399ff\") " Nov 22 05:23:39 crc kubenswrapper[4818]: I1122 05:23:39.707123 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/793ce69e-3676-4cdd-8b9a-97164a8399ff-utilities\") pod \"793ce69e-3676-4cdd-8b9a-97164a8399ff\" (UID: \"793ce69e-3676-4cdd-8b9a-97164a8399ff\") " Nov 22 05:23:39 crc kubenswrapper[4818]: I1122 05:23:39.708179 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/793ce69e-3676-4cdd-8b9a-97164a8399ff-utilities" (OuterVolumeSpecName: "utilities") pod "793ce69e-3676-4cdd-8b9a-97164a8399ff" (UID: "793ce69e-3676-4cdd-8b9a-97164a8399ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:23:39 crc kubenswrapper[4818]: I1122 05:23:39.711908 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/793ce69e-3676-4cdd-8b9a-97164a8399ff-kube-api-access-l27v9" (OuterVolumeSpecName: "kube-api-access-l27v9") pod "793ce69e-3676-4cdd-8b9a-97164a8399ff" (UID: "793ce69e-3676-4cdd-8b9a-97164a8399ff"). InnerVolumeSpecName "kube-api-access-l27v9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:23:39 crc kubenswrapper[4818]: I1122 05:23:39.790722 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/793ce69e-3676-4cdd-8b9a-97164a8399ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "793ce69e-3676-4cdd-8b9a-97164a8399ff" (UID: "793ce69e-3676-4cdd-8b9a-97164a8399ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:23:39 crc kubenswrapper[4818]: I1122 05:23:39.809413 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/793ce69e-3676-4cdd-8b9a-97164a8399ff-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 05:23:39 crc kubenswrapper[4818]: I1122 05:23:39.809467 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/793ce69e-3676-4cdd-8b9a-97164a8399ff-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 05:23:39 crc kubenswrapper[4818]: I1122 05:23:39.809514 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l27v9\" (UniqueName: \"kubernetes.io/projected/793ce69e-3676-4cdd-8b9a-97164a8399ff-kube-api-access-l27v9\") on node \"crc\" DevicePath \"\"" Nov 22 05:23:40 crc kubenswrapper[4818]: I1122 05:23:40.181405 4818 generic.go:334] "Generic (PLEG): container finished" podID="793ce69e-3676-4cdd-8b9a-97164a8399ff" containerID="80f6c106ab3fd9dbe8622602a0b326bc043a1727e585723b94f2f43c3078fd78" exitCode=0 Nov 22 05:23:40 crc kubenswrapper[4818]: I1122 05:23:40.181496 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lz4zd" Nov 22 05:23:40 crc kubenswrapper[4818]: I1122 05:23:40.181521 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lz4zd" event={"ID":"793ce69e-3676-4cdd-8b9a-97164a8399ff","Type":"ContainerDied","Data":"80f6c106ab3fd9dbe8622602a0b326bc043a1727e585723b94f2f43c3078fd78"} Nov 22 05:23:40 crc kubenswrapper[4818]: I1122 05:23:40.181908 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lz4zd" event={"ID":"793ce69e-3676-4cdd-8b9a-97164a8399ff","Type":"ContainerDied","Data":"ef7a88f8846a9640e960ab9d6a90e931fd0df46b96fcb49cb8c8d59755aeca85"} Nov 22 05:23:40 crc kubenswrapper[4818]: I1122 05:23:40.181931 4818 scope.go:117] "RemoveContainer" containerID="80f6c106ab3fd9dbe8622602a0b326bc043a1727e585723b94f2f43c3078fd78" Nov 22 05:23:40 crc kubenswrapper[4818]: I1122 05:23:40.200584 4818 scope.go:117] "RemoveContainer" containerID="64f66b41feda13b421a00a0947b33d8dbf8f07b1de0d50c0f0be7d21f641c7fe" Nov 22 05:23:40 crc kubenswrapper[4818]: I1122 05:23:40.227053 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lz4zd"] Nov 22 05:23:40 crc kubenswrapper[4818]: I1122 05:23:40.232029 4818 scope.go:117] "RemoveContainer" containerID="c83d54a48469030e65a2c53de994ca64e8bb882ad5e9dc4142e462eff87e491d" Nov 22 05:23:40 crc kubenswrapper[4818]: I1122 05:23:40.237853 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lz4zd"] Nov 22 05:23:40 crc kubenswrapper[4818]: I1122 05:23:40.251601 4818 scope.go:117] "RemoveContainer" containerID="80f6c106ab3fd9dbe8622602a0b326bc043a1727e585723b94f2f43c3078fd78" Nov 22 05:23:40 crc kubenswrapper[4818]: E1122 05:23:40.252045 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80f6c106ab3fd9dbe8622602a0b326bc043a1727e585723b94f2f43c3078fd78\": container with ID starting with 80f6c106ab3fd9dbe8622602a0b326bc043a1727e585723b94f2f43c3078fd78 not found: ID does not exist" containerID="80f6c106ab3fd9dbe8622602a0b326bc043a1727e585723b94f2f43c3078fd78" Nov 22 05:23:40 crc kubenswrapper[4818]: I1122 05:23:40.252079 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80f6c106ab3fd9dbe8622602a0b326bc043a1727e585723b94f2f43c3078fd78"} err="failed to get container status \"80f6c106ab3fd9dbe8622602a0b326bc043a1727e585723b94f2f43c3078fd78\": rpc error: code = NotFound desc = could not find container \"80f6c106ab3fd9dbe8622602a0b326bc043a1727e585723b94f2f43c3078fd78\": container with ID starting with 80f6c106ab3fd9dbe8622602a0b326bc043a1727e585723b94f2f43c3078fd78 not found: ID does not exist" Nov 22 05:23:40 crc kubenswrapper[4818]: I1122 05:23:40.252100 4818 scope.go:117] "RemoveContainer" containerID="64f66b41feda13b421a00a0947b33d8dbf8f07b1de0d50c0f0be7d21f641c7fe" Nov 22 05:23:40 crc kubenswrapper[4818]: E1122 05:23:40.252673 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64f66b41feda13b421a00a0947b33d8dbf8f07b1de0d50c0f0be7d21f641c7fe\": container with ID starting with 64f66b41feda13b421a00a0947b33d8dbf8f07b1de0d50c0f0be7d21f641c7fe not found: ID does not exist" containerID="64f66b41feda13b421a00a0947b33d8dbf8f07b1de0d50c0f0be7d21f641c7fe" Nov 22 05:23:40 crc kubenswrapper[4818]: I1122 05:23:40.252712 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64f66b41feda13b421a00a0947b33d8dbf8f07b1de0d50c0f0be7d21f641c7fe"} err="failed to get container status \"64f66b41feda13b421a00a0947b33d8dbf8f07b1de0d50c0f0be7d21f641c7fe\": rpc error: code = NotFound desc = could not find container \"64f66b41feda13b421a00a0947b33d8dbf8f07b1de0d50c0f0be7d21f641c7fe\": container with ID starting with 64f66b41feda13b421a00a0947b33d8dbf8f07b1de0d50c0f0be7d21f641c7fe not found: ID does not exist" Nov 22 05:23:40 crc kubenswrapper[4818]: I1122 05:23:40.252738 4818 scope.go:117] "RemoveContainer" containerID="c83d54a48469030e65a2c53de994ca64e8bb882ad5e9dc4142e462eff87e491d" Nov 22 05:23:40 crc kubenswrapper[4818]: E1122 05:23:40.253062 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c83d54a48469030e65a2c53de994ca64e8bb882ad5e9dc4142e462eff87e491d\": container with ID starting with c83d54a48469030e65a2c53de994ca64e8bb882ad5e9dc4142e462eff87e491d not found: ID does not exist" containerID="c83d54a48469030e65a2c53de994ca64e8bb882ad5e9dc4142e462eff87e491d" Nov 22 05:23:40 crc kubenswrapper[4818]: I1122 05:23:40.253092 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c83d54a48469030e65a2c53de994ca64e8bb882ad5e9dc4142e462eff87e491d"} err="failed to get container status \"c83d54a48469030e65a2c53de994ca64e8bb882ad5e9dc4142e462eff87e491d\": rpc error: code = NotFound desc = could not find container \"c83d54a48469030e65a2c53de994ca64e8bb882ad5e9dc4142e462eff87e491d\": container with ID starting with c83d54a48469030e65a2c53de994ca64e8bb882ad5e9dc4142e462eff87e491d not found: ID does not exist" Nov 22 05:23:40 crc kubenswrapper[4818]: I1122 05:23:40.309126 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="793ce69e-3676-4cdd-8b9a-97164a8399ff" path="/var/lib/kubelet/pods/793ce69e-3676-4cdd-8b9a-97164a8399ff/volumes" Nov 22 05:23:52 crc kubenswrapper[4818]: I1122 05:23:52.046488 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-lcvm7"] Nov 22 05:23:52 crc kubenswrapper[4818]: I1122 05:23:52.055537 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-lcvm7"] Nov 22 05:23:52 crc kubenswrapper[4818]: I1122 05:23:52.304218 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdfdf908-61b8-4dcd-9898-fafdee790ca0" path="/var/lib/kubelet/pods/cdfdf908-61b8-4dcd-9898-fafdee790ca0/volumes" Nov 22 05:24:39 crc kubenswrapper[4818]: I1122 05:24:39.757214 4818 scope.go:117] "RemoveContainer" containerID="42c56e4d11f063f9ca8ed3cfef725446df4c6f534b616e9dbc311be53c36d939" Nov 22 05:25:43 crc kubenswrapper[4818]: I1122 05:25:43.148186 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xvsnj"] Nov 22 05:25:43 crc kubenswrapper[4818]: E1122 05:25:43.150582 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="793ce69e-3676-4cdd-8b9a-97164a8399ff" containerName="extract-utilities" Nov 22 05:25:43 crc kubenswrapper[4818]: I1122 05:25:43.150605 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="793ce69e-3676-4cdd-8b9a-97164a8399ff" containerName="extract-utilities" Nov 22 05:25:43 crc kubenswrapper[4818]: E1122 05:25:43.150653 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="793ce69e-3676-4cdd-8b9a-97164a8399ff" containerName="extract-content" Nov 22 05:25:43 crc kubenswrapper[4818]: I1122 05:25:43.150660 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="793ce69e-3676-4cdd-8b9a-97164a8399ff" containerName="extract-content" Nov 22 05:25:43 crc kubenswrapper[4818]: E1122 05:25:43.150681 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="793ce69e-3676-4cdd-8b9a-97164a8399ff" containerName="registry-server" Nov 22 05:25:43 crc kubenswrapper[4818]: I1122 05:25:43.150688 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="793ce69e-3676-4cdd-8b9a-97164a8399ff" containerName="registry-server" Nov 22 05:25:43 crc kubenswrapper[4818]: I1122 05:25:43.150900 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="793ce69e-3676-4cdd-8b9a-97164a8399ff" containerName="registry-server" Nov 22 05:25:43 crc kubenswrapper[4818]: I1122 05:25:43.152674 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvsnj" Nov 22 05:25:43 crc kubenswrapper[4818]: I1122 05:25:43.171983 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xvsnj"] Nov 22 05:25:43 crc kubenswrapper[4818]: I1122 05:25:43.218027 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrsft\" (UniqueName: \"kubernetes.io/projected/95fe4c8b-10e5-4da6-a435-150df60b150c-kube-api-access-qrsft\") pod \"community-operators-xvsnj\" (UID: \"95fe4c8b-10e5-4da6-a435-150df60b150c\") " pod="openshift-marketplace/community-operators-xvsnj" Nov 22 05:25:43 crc kubenswrapper[4818]: I1122 05:25:43.218153 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95fe4c8b-10e5-4da6-a435-150df60b150c-utilities\") pod \"community-operators-xvsnj\" (UID: \"95fe4c8b-10e5-4da6-a435-150df60b150c\") " pod="openshift-marketplace/community-operators-xvsnj" Nov 22 05:25:43 crc kubenswrapper[4818]: I1122 05:25:43.218186 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95fe4c8b-10e5-4da6-a435-150df60b150c-catalog-content\") pod \"community-operators-xvsnj\" (UID: \"95fe4c8b-10e5-4da6-a435-150df60b150c\") " pod="openshift-marketplace/community-operators-xvsnj" Nov 22 05:25:43 crc kubenswrapper[4818]: I1122 05:25:43.319867 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrsft\" (UniqueName: \"kubernetes.io/projected/95fe4c8b-10e5-4da6-a435-150df60b150c-kube-api-access-qrsft\") pod \"community-operators-xvsnj\" (UID: \"95fe4c8b-10e5-4da6-a435-150df60b150c\") " pod="openshift-marketplace/community-operators-xvsnj" Nov 22 05:25:43 crc kubenswrapper[4818]: I1122 05:25:43.319961 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95fe4c8b-10e5-4da6-a435-150df60b150c-utilities\") pod \"community-operators-xvsnj\" (UID: \"95fe4c8b-10e5-4da6-a435-150df60b150c\") " pod="openshift-marketplace/community-operators-xvsnj" Nov 22 05:25:43 crc kubenswrapper[4818]: I1122 05:25:43.319995 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95fe4c8b-10e5-4da6-a435-150df60b150c-catalog-content\") pod \"community-operators-xvsnj\" (UID: \"95fe4c8b-10e5-4da6-a435-150df60b150c\") " pod="openshift-marketplace/community-operators-xvsnj" Nov 22 05:25:43 crc kubenswrapper[4818]: I1122 05:25:43.320540 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95fe4c8b-10e5-4da6-a435-150df60b150c-utilities\") pod \"community-operators-xvsnj\" (UID: \"95fe4c8b-10e5-4da6-a435-150df60b150c\") " pod="openshift-marketplace/community-operators-xvsnj" Nov 22 05:25:43 crc kubenswrapper[4818]: I1122 05:25:43.320642 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95fe4c8b-10e5-4da6-a435-150df60b150c-catalog-content\") pod \"community-operators-xvsnj\" (UID: \"95fe4c8b-10e5-4da6-a435-150df60b150c\") " pod="openshift-marketplace/community-operators-xvsnj" Nov 22 05:25:43 crc kubenswrapper[4818]: I1122 05:25:43.340424 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrsft\" (UniqueName: \"kubernetes.io/projected/95fe4c8b-10e5-4da6-a435-150df60b150c-kube-api-access-qrsft\") pod \"community-operators-xvsnj\" (UID: \"95fe4c8b-10e5-4da6-a435-150df60b150c\") " pod="openshift-marketplace/community-operators-xvsnj" Nov 22 05:25:43 crc kubenswrapper[4818]: I1122 05:25:43.532402 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvsnj" Nov 22 05:25:43 crc kubenswrapper[4818]: I1122 05:25:43.987632 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xvsnj"] Nov 22 05:25:44 crc kubenswrapper[4818]: I1122 05:25:44.518578 4818 generic.go:334] "Generic (PLEG): container finished" podID="95fe4c8b-10e5-4da6-a435-150df60b150c" containerID="1e92f60519d327fc048636ef50e60df5a77a1edb2c20805ed6990b151838e2b6" exitCode=0 Nov 22 05:25:44 crc kubenswrapper[4818]: I1122 05:25:44.518695 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvsnj" event={"ID":"95fe4c8b-10e5-4da6-a435-150df60b150c","Type":"ContainerDied","Data":"1e92f60519d327fc048636ef50e60df5a77a1edb2c20805ed6990b151838e2b6"} Nov 22 05:25:44 crc kubenswrapper[4818]: I1122 05:25:44.519016 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvsnj" event={"ID":"95fe4c8b-10e5-4da6-a435-150df60b150c","Type":"ContainerStarted","Data":"c871874f144b706b0ea8b9762b6c23d1113446e7f2114e30cd26a6a5747cae0b"} Nov 22 05:25:44 crc kubenswrapper[4818]: I1122 05:25:44.522125 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 05:25:45 crc kubenswrapper[4818]: I1122 05:25:45.531658 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvsnj" event={"ID":"95fe4c8b-10e5-4da6-a435-150df60b150c","Type":"ContainerStarted","Data":"53124be5550c9dda76d01dd6b74f89106aed392fa709e5bead0f9382222c9d52"} Nov 22 05:25:46 crc kubenswrapper[4818]: I1122 05:25:46.552541 4818 generic.go:334] "Generic (PLEG): container finished" podID="95fe4c8b-10e5-4da6-a435-150df60b150c" containerID="53124be5550c9dda76d01dd6b74f89106aed392fa709e5bead0f9382222c9d52" exitCode=0 Nov 22 05:25:46 crc kubenswrapper[4818]: I1122 05:25:46.552658 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvsnj" event={"ID":"95fe4c8b-10e5-4da6-a435-150df60b150c","Type":"ContainerDied","Data":"53124be5550c9dda76d01dd6b74f89106aed392fa709e5bead0f9382222c9d52"} Nov 22 05:25:47 crc kubenswrapper[4818]: I1122 05:25:47.563536 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvsnj" event={"ID":"95fe4c8b-10e5-4da6-a435-150df60b150c","Type":"ContainerStarted","Data":"53c7f943a7c243b17091fdf730a257cfd57e7730b09a468954aab44e79ddc508"} Nov 22 05:25:47 crc kubenswrapper[4818]: I1122 05:25:47.590003 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xvsnj" podStartSLOduration=1.975072744 podStartE2EDuration="4.58997815s" podCreationTimestamp="2025-11-22 05:25:43 +0000 UTC" firstStartedPulling="2025-11-22 05:25:44.521653754 +0000 UTC m=+2297.096070311" lastFinishedPulling="2025-11-22 05:25:47.13655914 +0000 UTC m=+2299.710975717" observedRunningTime="2025-11-22 05:25:47.58401814 +0000 UTC m=+2300.158434707" watchObservedRunningTime="2025-11-22 05:25:47.58997815 +0000 UTC m=+2300.164394677" Nov 22 05:25:51 crc kubenswrapper[4818]: I1122 05:25:51.264536 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:25:51 crc kubenswrapper[4818]: I1122 05:25:51.264914 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:25:53 crc kubenswrapper[4818]: I1122 05:25:53.533448 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xvsnj" Nov 22 05:25:53 crc kubenswrapper[4818]: I1122 05:25:53.534056 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xvsnj" Nov 22 05:25:53 crc kubenswrapper[4818]: I1122 05:25:53.589644 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xvsnj" Nov 22 05:25:53 crc kubenswrapper[4818]: I1122 05:25:53.663924 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xvsnj" Nov 22 05:25:53 crc kubenswrapper[4818]: I1122 05:25:53.825885 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xvsnj"] Nov 22 05:25:55 crc kubenswrapper[4818]: I1122 05:25:55.642588 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xvsnj" podUID="95fe4c8b-10e5-4da6-a435-150df60b150c" containerName="registry-server" containerID="cri-o://53c7f943a7c243b17091fdf730a257cfd57e7730b09a468954aab44e79ddc508" gracePeriod=2 Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.272530 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvsnj" Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.288702 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95fe4c8b-10e5-4da6-a435-150df60b150c-catalog-content\") pod \"95fe4c8b-10e5-4da6-a435-150df60b150c\" (UID: \"95fe4c8b-10e5-4da6-a435-150df60b150c\") " Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.290084 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrsft\" (UniqueName: \"kubernetes.io/projected/95fe4c8b-10e5-4da6-a435-150df60b150c-kube-api-access-qrsft\") pod \"95fe4c8b-10e5-4da6-a435-150df60b150c\" (UID: \"95fe4c8b-10e5-4da6-a435-150df60b150c\") " Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.290235 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95fe4c8b-10e5-4da6-a435-150df60b150c-utilities\") pod \"95fe4c8b-10e5-4da6-a435-150df60b150c\" (UID: \"95fe4c8b-10e5-4da6-a435-150df60b150c\") " Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.296090 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95fe4c8b-10e5-4da6-a435-150df60b150c-utilities" (OuterVolumeSpecName: "utilities") pod "95fe4c8b-10e5-4da6-a435-150df60b150c" (UID: "95fe4c8b-10e5-4da6-a435-150df60b150c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.304547 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95fe4c8b-10e5-4da6-a435-150df60b150c-kube-api-access-qrsft" (OuterVolumeSpecName: "kube-api-access-qrsft") pod "95fe4c8b-10e5-4da6-a435-150df60b150c" (UID: "95fe4c8b-10e5-4da6-a435-150df60b150c"). InnerVolumeSpecName "kube-api-access-qrsft". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.395125 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrsft\" (UniqueName: \"kubernetes.io/projected/95fe4c8b-10e5-4da6-a435-150df60b150c-kube-api-access-qrsft\") on node \"crc\" DevicePath \"\"" Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.395229 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95fe4c8b-10e5-4da6-a435-150df60b150c-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.660112 4818 generic.go:334] "Generic (PLEG): container finished" podID="95fe4c8b-10e5-4da6-a435-150df60b150c" containerID="53c7f943a7c243b17091fdf730a257cfd57e7730b09a468954aab44e79ddc508" exitCode=0 Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.660175 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvsnj" Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.660167 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvsnj" event={"ID":"95fe4c8b-10e5-4da6-a435-150df60b150c","Type":"ContainerDied","Data":"53c7f943a7c243b17091fdf730a257cfd57e7730b09a468954aab44e79ddc508"} Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.660659 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvsnj" event={"ID":"95fe4c8b-10e5-4da6-a435-150df60b150c","Type":"ContainerDied","Data":"c871874f144b706b0ea8b9762b6c23d1113446e7f2114e30cd26a6a5747cae0b"} Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.660687 4818 scope.go:117] "RemoveContainer" containerID="53c7f943a7c243b17091fdf730a257cfd57e7730b09a468954aab44e79ddc508" Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.682669 4818 scope.go:117] "RemoveContainer" containerID="53124be5550c9dda76d01dd6b74f89106aed392fa709e5bead0f9382222c9d52" Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.712787 4818 scope.go:117] "RemoveContainer" containerID="1e92f60519d327fc048636ef50e60df5a77a1edb2c20805ed6990b151838e2b6" Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.760095 4818 scope.go:117] "RemoveContainer" containerID="53c7f943a7c243b17091fdf730a257cfd57e7730b09a468954aab44e79ddc508" Nov 22 05:25:56 crc kubenswrapper[4818]: E1122 05:25:56.760649 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53c7f943a7c243b17091fdf730a257cfd57e7730b09a468954aab44e79ddc508\": container with ID starting with 53c7f943a7c243b17091fdf730a257cfd57e7730b09a468954aab44e79ddc508 not found: ID does not exist" containerID="53c7f943a7c243b17091fdf730a257cfd57e7730b09a468954aab44e79ddc508" Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.760708 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53c7f943a7c243b17091fdf730a257cfd57e7730b09a468954aab44e79ddc508"} err="failed to get container status \"53c7f943a7c243b17091fdf730a257cfd57e7730b09a468954aab44e79ddc508\": rpc error: code = NotFound desc = could not find container \"53c7f943a7c243b17091fdf730a257cfd57e7730b09a468954aab44e79ddc508\": container with ID starting with 53c7f943a7c243b17091fdf730a257cfd57e7730b09a468954aab44e79ddc508 not found: ID does not exist" Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.760745 4818 scope.go:117] "RemoveContainer" containerID="53124be5550c9dda76d01dd6b74f89106aed392fa709e5bead0f9382222c9d52" Nov 22 05:25:56 crc kubenswrapper[4818]: E1122 05:25:56.761166 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53124be5550c9dda76d01dd6b74f89106aed392fa709e5bead0f9382222c9d52\": container with ID starting with 53124be5550c9dda76d01dd6b74f89106aed392fa709e5bead0f9382222c9d52 not found: ID does not exist" containerID="53124be5550c9dda76d01dd6b74f89106aed392fa709e5bead0f9382222c9d52" Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.761211 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53124be5550c9dda76d01dd6b74f89106aed392fa709e5bead0f9382222c9d52"} err="failed to get container status \"53124be5550c9dda76d01dd6b74f89106aed392fa709e5bead0f9382222c9d52\": rpc error: code = NotFound desc = could not find container \"53124be5550c9dda76d01dd6b74f89106aed392fa709e5bead0f9382222c9d52\": container with ID starting with 53124be5550c9dda76d01dd6b74f89106aed392fa709e5bead0f9382222c9d52 not found: ID does not exist" Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.761236 4818 scope.go:117] "RemoveContainer" containerID="1e92f60519d327fc048636ef50e60df5a77a1edb2c20805ed6990b151838e2b6" Nov 22 05:25:56 crc kubenswrapper[4818]: E1122 05:25:56.761628 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e92f60519d327fc048636ef50e60df5a77a1edb2c20805ed6990b151838e2b6\": container with ID starting with 1e92f60519d327fc048636ef50e60df5a77a1edb2c20805ed6990b151838e2b6 not found: ID does not exist" containerID="1e92f60519d327fc048636ef50e60df5a77a1edb2c20805ed6990b151838e2b6" Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.761660 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e92f60519d327fc048636ef50e60df5a77a1edb2c20805ed6990b151838e2b6"} err="failed to get container status \"1e92f60519d327fc048636ef50e60df5a77a1edb2c20805ed6990b151838e2b6\": rpc error: code = NotFound desc = could not find container \"1e92f60519d327fc048636ef50e60df5a77a1edb2c20805ed6990b151838e2b6\": container with ID starting with 1e92f60519d327fc048636ef50e60df5a77a1edb2c20805ed6990b151838e2b6 not found: ID does not exist" Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.809551 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95fe4c8b-10e5-4da6-a435-150df60b150c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95fe4c8b-10e5-4da6-a435-150df60b150c" (UID: "95fe4c8b-10e5-4da6-a435-150df60b150c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.904811 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95fe4c8b-10e5-4da6-a435-150df60b150c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.991201 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xvsnj"] Nov 22 05:25:56 crc kubenswrapper[4818]: I1122 05:25:56.997895 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xvsnj"] Nov 22 05:25:58 crc kubenswrapper[4818]: I1122 05:25:58.312200 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95fe4c8b-10e5-4da6-a435-150df60b150c" path="/var/lib/kubelet/pods/95fe4c8b-10e5-4da6-a435-150df60b150c/volumes" Nov 22 05:26:21 crc kubenswrapper[4818]: I1122 05:26:21.265564 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:26:21 crc kubenswrapper[4818]: I1122 05:26:21.266360 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:26:51 crc kubenswrapper[4818]: I1122 05:26:51.264562 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:26:51 crc kubenswrapper[4818]: I1122 05:26:51.264991 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:26:51 crc kubenswrapper[4818]: I1122 05:26:51.265045 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 05:26:51 crc kubenswrapper[4818]: I1122 05:26:51.266050 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc"} pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 05:26:51 crc kubenswrapper[4818]: I1122 05:26:51.266103 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" containerID="cri-o://0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" gracePeriod=600 Nov 22 05:26:51 crc kubenswrapper[4818]: E1122 05:26:51.403599 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:26:52 crc kubenswrapper[4818]: I1122 05:26:52.184970 4818 generic.go:334] "Generic (PLEG): container finished" podID="0971c9ec-29ce-44eb-9875-ad4664143305" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" exitCode=0 Nov 22 05:26:52 crc kubenswrapper[4818]: I1122 05:26:52.185052 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerDied","Data":"0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc"} Nov 22 05:26:52 crc kubenswrapper[4818]: I1122 05:26:52.185386 4818 scope.go:117] "RemoveContainer" containerID="d3ddef518d72589af006b70b271591f843cb53e6b99204dda0a3b0fa65c54eb9" Nov 22 05:26:52 crc kubenswrapper[4818]: I1122 05:26:52.186352 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:26:52 crc kubenswrapper[4818]: E1122 05:26:52.186892 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:27:06 crc kubenswrapper[4818]: I1122 05:27:06.291501 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:27:06 crc kubenswrapper[4818]: E1122 05:27:06.292925 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:27:21 crc kubenswrapper[4818]: I1122 05:27:21.292772 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:27:21 crc kubenswrapper[4818]: E1122 05:27:21.295019 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:27:31 crc kubenswrapper[4818]: E1122 05:27:31.275130 4818 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.110:53248->38.102.83.110:39825: write tcp 38.102.83.110:53248->38.102.83.110:39825: write: connection reset by peer Nov 22 05:27:33 crc kubenswrapper[4818]: I1122 05:27:33.292098 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:27:33 crc kubenswrapper[4818]: E1122 05:27:33.292906 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:27:40 crc kubenswrapper[4818]: E1122 05:27:40.564946 4818 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.110:59066->38.102.83.110:39825: write tcp 38.102.83.110:59066->38.102.83.110:39825: write: connection reset by peer Nov 22 05:27:45 crc kubenswrapper[4818]: E1122 05:27:45.211857 4818 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.110:59100->38.102.83.110:39825: read tcp 38.102.83.110:59100->38.102.83.110:39825: read: connection reset by peer Nov 22 05:27:45 crc kubenswrapper[4818]: E1122 05:27:45.212445 4818 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.110:59100->38.102.83.110:39825: write tcp 38.102.83.110:59100->38.102.83.110:39825: write: broken pipe Nov 22 05:27:46 crc kubenswrapper[4818]: I1122 05:27:46.290994 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:27:46 crc kubenswrapper[4818]: E1122 05:27:46.291264 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:27:57 crc kubenswrapper[4818]: I1122 05:27:57.293301 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:27:57 crc kubenswrapper[4818]: E1122 05:27:57.294730 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.016016 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv"] Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.023091 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49cv"] Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.034077 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m"] Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.043305 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk"] Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.050130 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-b85b7"] Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.056511 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hdr5m"] Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.063601 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6"] Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.069711 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw"] Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.077078 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h"] Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.083477 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt"] Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.090359 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p"] Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.096107 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc"] Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.102010 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vd8mw"] Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.133155 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-kczgk"] Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.140787 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-b85b7"] Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.147424 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ml85h"] Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.153462 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-p956p"] Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.159885 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-87jlc"] Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.166564 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6zlt6"] Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.173393 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5cxt"] Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.329053 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ffc9da3-3fd2-4752-9d8d-f22e149e375d" path="/var/lib/kubelet/pods/2ffc9da3-3fd2-4752-9d8d-f22e149e375d/volumes" Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.329842 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46a98f54-df60-48ae-a6ee-5affd771243e" path="/var/lib/kubelet/pods/46a98f54-df60-48ae-a6ee-5affd771243e/volumes" Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.330538 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79833bd6-1636-441a-8fae-e319a7d0873e" path="/var/lib/kubelet/pods/79833bd6-1636-441a-8fae-e319a7d0873e/volumes" Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.331244 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ce7c513-156b-4380-a2da-1a2ab6ee32e4" path="/var/lib/kubelet/pods/7ce7c513-156b-4380-a2da-1a2ab6ee32e4/volumes" Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.332196 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9461206a-81ad-4c30-a0fe-18ecf0502570" path="/var/lib/kubelet/pods/9461206a-81ad-4c30-a0fe-18ecf0502570/volumes" Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.332697 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfcce923-6850-4c78-b11f-e71ae4c680db" path="/var/lib/kubelet/pods/bfcce923-6850-4c78-b11f-e71ae4c680db/volumes" Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.333232 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d49504af-1b9b-4f48-99b1-7c5ef020f27c" path="/var/lib/kubelet/pods/d49504af-1b9b-4f48-99b1-7c5ef020f27c/volumes" Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.334155 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc7ffd42-a7db-44b2-851c-1a074bb23eb7" path="/var/lib/kubelet/pods/dc7ffd42-a7db-44b2-851c-1a074bb23eb7/volumes" Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.334652 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f362e5c8-ad07-472c-93f2-acf126d4135a" path="/var/lib/kubelet/pods/f362e5c8-ad07-472c-93f2-acf126d4135a/volumes" Nov 22 05:28:06 crc kubenswrapper[4818]: I1122 05:28:06.335131 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fee6592f-65ea-480e-813a-b4839af8fb95" path="/var/lib/kubelet/pods/fee6592f-65ea-480e-813a-b4839af8fb95/volumes" Nov 22 05:28:11 crc kubenswrapper[4818]: I1122 05:28:11.292389 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:28:11 crc kubenswrapper[4818]: E1122 05:28:11.293575 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.170366 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst"] Nov 22 05:28:12 crc kubenswrapper[4818]: E1122 05:28:12.170886 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95fe4c8b-10e5-4da6-a435-150df60b150c" containerName="extract-content" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.170912 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="95fe4c8b-10e5-4da6-a435-150df60b150c" containerName="extract-content" Nov 22 05:28:12 crc kubenswrapper[4818]: E1122 05:28:12.170923 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95fe4c8b-10e5-4da6-a435-150df60b150c" containerName="registry-server" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.170929 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="95fe4c8b-10e5-4da6-a435-150df60b150c" containerName="registry-server" Nov 22 05:28:12 crc kubenswrapper[4818]: E1122 05:28:12.170957 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95fe4c8b-10e5-4da6-a435-150df60b150c" containerName="extract-utilities" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.170964 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="95fe4c8b-10e5-4da6-a435-150df60b150c" containerName="extract-utilities" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.171124 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="95fe4c8b-10e5-4da6-a435-150df60b150c" containerName="registry-server" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.171729 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.174310 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.174420 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.174655 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.174687 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.175006 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.190179 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst"] Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.285494 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst\" (UID: \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.285838 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst\" (UID: \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.285906 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr6lc\" (UniqueName: \"kubernetes.io/projected/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-kube-api-access-wr6lc\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst\" (UID: \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.286075 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst\" (UID: \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.286371 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst\" (UID: \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.387861 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst\" (UID: \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.387931 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst\" (UID: \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.387973 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst\" (UID: \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.388127 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst\" (UID: \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.388157 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr6lc\" (UniqueName: \"kubernetes.io/projected/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-kube-api-access-wr6lc\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst\" (UID: \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.397681 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst\" (UID: \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.398122 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst\" (UID: \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.399629 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst\" (UID: \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.403021 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst\" (UID: \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.410859 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr6lc\" (UniqueName: \"kubernetes.io/projected/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-kube-api-access-wr6lc\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst\" (UID: \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" Nov 22 05:28:12 crc kubenswrapper[4818]: I1122 05:28:12.489222 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" Nov 22 05:28:13 crc kubenswrapper[4818]: I1122 05:28:13.092117 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst"] Nov 22 05:28:14 crc kubenswrapper[4818]: I1122 05:28:14.026327 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" event={"ID":"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2","Type":"ContainerStarted","Data":"6bbde5e2f3f08c6806d5e3c1ab41fc02479c3a2afeddf219e65d493f024843ac"} Nov 22 05:28:14 crc kubenswrapper[4818]: I1122 05:28:14.027022 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" event={"ID":"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2","Type":"ContainerStarted","Data":"d2c70b8ab6bc4f9569dd985766471a0e0e8bc8271021264224d08e95cc506d6e"} Nov 22 05:28:14 crc kubenswrapper[4818]: I1122 05:28:14.056955 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" podStartSLOduration=1.600995049 podStartE2EDuration="2.056929745s" podCreationTimestamp="2025-11-22 05:28:12 +0000 UTC" firstStartedPulling="2025-11-22 05:28:13.112206508 +0000 UTC m=+2445.686623045" lastFinishedPulling="2025-11-22 05:28:13.568141174 +0000 UTC m=+2446.142557741" observedRunningTime="2025-11-22 05:28:14.048410195 +0000 UTC m=+2446.622826732" watchObservedRunningTime="2025-11-22 05:28:14.056929745 +0000 UTC m=+2446.631346282" Nov 22 05:28:22 crc kubenswrapper[4818]: I1122 05:28:22.292137 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:28:22 crc kubenswrapper[4818]: E1122 05:28:22.293178 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:28:25 crc kubenswrapper[4818]: E1122 05:28:25.524087 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c466daa_c1c7_4d3c_a502_cfb8f424f6b2.slice/crio-conmon-6bbde5e2f3f08c6806d5e3c1ab41fc02479c3a2afeddf219e65d493f024843ac.scope\": RecentStats: unable to find data in memory cache]" Nov 22 05:28:26 crc kubenswrapper[4818]: I1122 05:28:26.141534 4818 generic.go:334] "Generic (PLEG): container finished" podID="0c466daa-c1c7-4d3c-a502-cfb8f424f6b2" containerID="6bbde5e2f3f08c6806d5e3c1ab41fc02479c3a2afeddf219e65d493f024843ac" exitCode=0 Nov 22 05:28:26 crc kubenswrapper[4818]: I1122 05:28:26.141589 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" event={"ID":"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2","Type":"ContainerDied","Data":"6bbde5e2f3f08c6806d5e3c1ab41fc02479c3a2afeddf219e65d493f024843ac"} Nov 22 05:28:27 crc kubenswrapper[4818]: I1122 05:28:27.546321 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" Nov 22 05:28:27 crc kubenswrapper[4818]: I1122 05:28:27.586477 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wr6lc\" (UniqueName: \"kubernetes.io/projected/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-kube-api-access-wr6lc\") pod \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\" (UID: \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\") " Nov 22 05:28:27 crc kubenswrapper[4818]: I1122 05:28:27.586621 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-inventory\") pod \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\" (UID: \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\") " Nov 22 05:28:27 crc kubenswrapper[4818]: I1122 05:28:27.586682 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-ssh-key\") pod \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\" (UID: \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\") " Nov 22 05:28:27 crc kubenswrapper[4818]: I1122 05:28:27.586784 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-repo-setup-combined-ca-bundle\") pod \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\" (UID: \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\") " Nov 22 05:28:27 crc kubenswrapper[4818]: I1122 05:28:27.586849 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-ceph\") pod \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\" (UID: \"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2\") " Nov 22 05:28:27 crc kubenswrapper[4818]: I1122 05:28:27.592992 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-ceph" (OuterVolumeSpecName: "ceph") pod "0c466daa-c1c7-4d3c-a502-cfb8f424f6b2" (UID: "0c466daa-c1c7-4d3c-a502-cfb8f424f6b2"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:28:27 crc kubenswrapper[4818]: I1122 05:28:27.593812 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-kube-api-access-wr6lc" (OuterVolumeSpecName: "kube-api-access-wr6lc") pod "0c466daa-c1c7-4d3c-a502-cfb8f424f6b2" (UID: "0c466daa-c1c7-4d3c-a502-cfb8f424f6b2"). InnerVolumeSpecName "kube-api-access-wr6lc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:28:27 crc kubenswrapper[4818]: I1122 05:28:27.593836 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "0c466daa-c1c7-4d3c-a502-cfb8f424f6b2" (UID: "0c466daa-c1c7-4d3c-a502-cfb8f424f6b2"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:28:27 crc kubenswrapper[4818]: I1122 05:28:27.615656 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0c466daa-c1c7-4d3c-a502-cfb8f424f6b2" (UID: "0c466daa-c1c7-4d3c-a502-cfb8f424f6b2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:28:27 crc kubenswrapper[4818]: I1122 05:28:27.627965 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-inventory" (OuterVolumeSpecName: "inventory") pod "0c466daa-c1c7-4d3c-a502-cfb8f424f6b2" (UID: "0c466daa-c1c7-4d3c-a502-cfb8f424f6b2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:28:27 crc kubenswrapper[4818]: I1122 05:28:27.689055 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wr6lc\" (UniqueName: \"kubernetes.io/projected/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-kube-api-access-wr6lc\") on node \"crc\" DevicePath \"\"" Nov 22 05:28:27 crc kubenswrapper[4818]: I1122 05:28:27.689101 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:28:27 crc kubenswrapper[4818]: I1122 05:28:27.689117 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:28:27 crc kubenswrapper[4818]: I1122 05:28:27.689129 4818 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:28:27 crc kubenswrapper[4818]: I1122 05:28:27.689144 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0c466daa-c1c7-4d3c-a502-cfb8f424f6b2-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.159049 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" event={"ID":"0c466daa-c1c7-4d3c-a502-cfb8f424f6b2","Type":"ContainerDied","Data":"d2c70b8ab6bc4f9569dd985766471a0e0e8bc8271021264224d08e95cc506d6e"} Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.159094 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2c70b8ab6bc4f9569dd985766471a0e0e8bc8271021264224d08e95cc506d6e" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.159092 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.253017 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g"] Nov 22 05:28:28 crc kubenswrapper[4818]: E1122 05:28:28.253390 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c466daa-c1c7-4d3c-a502-cfb8f424f6b2" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.253412 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c466daa-c1c7-4d3c-a502-cfb8f424f6b2" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.253631 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c466daa-c1c7-4d3c-a502-cfb8f424f6b2" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.254175 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.256242 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.256785 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.257097 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.257354 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.257681 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.272771 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g"] Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.400983 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g\" (UID: \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.401468 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbx5r\" (UniqueName: \"kubernetes.io/projected/597bc073-9fa3-485c-b23d-8f7b5ea21da9-kube-api-access-vbx5r\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g\" (UID: \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.401566 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g\" (UID: \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.401598 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g\" (UID: \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.402056 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g\" (UID: \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.504281 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g\" (UID: \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.504413 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g\" (UID: \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.504437 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbx5r\" (UniqueName: \"kubernetes.io/projected/597bc073-9fa3-485c-b23d-8f7b5ea21da9-kube-api-access-vbx5r\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g\" (UID: \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.504491 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g\" (UID: \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.504529 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g\" (UID: \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.508737 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g\" (UID: \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.508995 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g\" (UID: \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.509869 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g\" (UID: \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.511715 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g\" (UID: \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.528861 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbx5r\" (UniqueName: \"kubernetes.io/projected/597bc073-9fa3-485c-b23d-8f7b5ea21da9-kube-api-access-vbx5r\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g\" (UID: \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" Nov 22 05:28:28 crc kubenswrapper[4818]: I1122 05:28:28.575897 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" Nov 22 05:28:29 crc kubenswrapper[4818]: I1122 05:28:29.148740 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g"] Nov 22 05:28:29 crc kubenswrapper[4818]: W1122 05:28:29.155936 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod597bc073_9fa3_485c_b23d_8f7b5ea21da9.slice/crio-dbcc23b10c35186df0b90c73eef5b12989d7529146a08ad260c20571c4aae0dc WatchSource:0}: Error finding container dbcc23b10c35186df0b90c73eef5b12989d7529146a08ad260c20571c4aae0dc: Status 404 returned error can't find the container with id dbcc23b10c35186df0b90c73eef5b12989d7529146a08ad260c20571c4aae0dc Nov 22 05:28:29 crc kubenswrapper[4818]: I1122 05:28:29.178425 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" event={"ID":"597bc073-9fa3-485c-b23d-8f7b5ea21da9","Type":"ContainerStarted","Data":"dbcc23b10c35186df0b90c73eef5b12989d7529146a08ad260c20571c4aae0dc"} Nov 22 05:28:30 crc kubenswrapper[4818]: I1122 05:28:30.191139 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" event={"ID":"597bc073-9fa3-485c-b23d-8f7b5ea21da9","Type":"ContainerStarted","Data":"74281e6854517e3d04a170a5ccbc81b5b2ac60259a5492e0a70a7c3f555a0fed"} Nov 22 05:28:30 crc kubenswrapper[4818]: I1122 05:28:30.206648 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" podStartSLOduration=1.7642532260000001 podStartE2EDuration="2.206630117s" podCreationTimestamp="2025-11-22 05:28:28 +0000 UTC" firstStartedPulling="2025-11-22 05:28:29.159303478 +0000 UTC m=+2461.733720015" lastFinishedPulling="2025-11-22 05:28:29.601680339 +0000 UTC m=+2462.176096906" observedRunningTime="2025-11-22 05:28:30.205477337 +0000 UTC m=+2462.779893874" watchObservedRunningTime="2025-11-22 05:28:30.206630117 +0000 UTC m=+2462.781046644" Nov 22 05:28:36 crc kubenswrapper[4818]: I1122 05:28:36.291654 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:28:36 crc kubenswrapper[4818]: E1122 05:28:36.292955 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:28:39 crc kubenswrapper[4818]: I1122 05:28:39.922143 4818 scope.go:117] "RemoveContainer" containerID="145cc5b5499a5258289402b4d7517e67fb6c9c1cc9c7540b0d591c98cf0f1b68" Nov 22 05:28:39 crc kubenswrapper[4818]: I1122 05:28:39.966373 4818 scope.go:117] "RemoveContainer" containerID="b3c0a62a2d03a49c5a08b84ffab21d422ca50344ad57d515a03841ecfd5036a7" Nov 22 05:28:40 crc kubenswrapper[4818]: I1122 05:28:40.028118 4818 scope.go:117] "RemoveContainer" containerID="f69d404bd56497fa60db8a5d2d8d155570aff7bd346fa1bd1792bac0d08c2a0c" Nov 22 05:28:40 crc kubenswrapper[4818]: I1122 05:28:40.056243 4818 scope.go:117] "RemoveContainer" containerID="d8ae2097cf9d1f5e554d2d06c50af3b6906b2d498b8ad631c02ced6c42434b39" Nov 22 05:28:40 crc kubenswrapper[4818]: I1122 05:28:40.114999 4818 scope.go:117] "RemoveContainer" containerID="5e903e3e975b2786aa56ea33b7ead03f01d47ccf166695a4359248206c7aada5" Nov 22 05:28:40 crc kubenswrapper[4818]: I1122 05:28:40.146765 4818 scope.go:117] "RemoveContainer" containerID="342208bbd7228f9f32a772585b9ad640b0713c4700a26825ae4946a4c5276f53" Nov 22 05:28:40 crc kubenswrapper[4818]: I1122 05:28:40.235853 4818 scope.go:117] "RemoveContainer" containerID="e2358be6b51f69e9623a757a076795204004cc0647800af27c0c6b26f0544379" Nov 22 05:28:40 crc kubenswrapper[4818]: I1122 05:28:40.272524 4818 scope.go:117] "RemoveContainer" containerID="5540a977d6599414b040c7baccf23573d7220dc64daca33947a9bfc7ab4b08aa" Nov 22 05:28:40 crc kubenswrapper[4818]: I1122 05:28:40.333162 4818 scope.go:117] "RemoveContainer" containerID="c95b0ec7b59f1472d7b7b8f88d063226e609dd63781d5ab875d78d7c5197480e" Nov 22 05:28:48 crc kubenswrapper[4818]: I1122 05:28:48.297028 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:28:48 crc kubenswrapper[4818]: E1122 05:28:48.298021 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:28:59 crc kubenswrapper[4818]: I1122 05:28:59.293878 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:28:59 crc kubenswrapper[4818]: E1122 05:28:59.294817 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:29:10 crc kubenswrapper[4818]: I1122 05:29:10.291033 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:29:10 crc kubenswrapper[4818]: E1122 05:29:10.291741 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:29:25 crc kubenswrapper[4818]: I1122 05:29:25.292135 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:29:25 crc kubenswrapper[4818]: E1122 05:29:25.293241 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:29:39 crc kubenswrapper[4818]: I1122 05:29:39.291053 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:29:39 crc kubenswrapper[4818]: E1122 05:29:39.291838 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:29:40 crc kubenswrapper[4818]: I1122 05:29:40.568644 4818 scope.go:117] "RemoveContainer" containerID="afe7820a66f248564a7448379bba8a9560929ca8857b3a4606b77acaa5938d27" Nov 22 05:29:52 crc kubenswrapper[4818]: I1122 05:29:52.292103 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:29:52 crc kubenswrapper[4818]: E1122 05:29:52.293320 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:30:00 crc kubenswrapper[4818]: I1122 05:30:00.154373 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls"] Nov 22 05:30:00 crc kubenswrapper[4818]: I1122 05:30:00.157679 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls" Nov 22 05:30:00 crc kubenswrapper[4818]: I1122 05:30:00.159988 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 05:30:00 crc kubenswrapper[4818]: I1122 05:30:00.161715 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 05:30:00 crc kubenswrapper[4818]: I1122 05:30:00.166853 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls"] Nov 22 05:30:00 crc kubenswrapper[4818]: I1122 05:30:00.286733 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5c0d4c12-ebd6-4020-aa27-d5dfda947b44-secret-volume\") pod \"collect-profiles-29396490-2ksls\" (UID: \"5c0d4c12-ebd6-4020-aa27-d5dfda947b44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls" Nov 22 05:30:00 crc kubenswrapper[4818]: I1122 05:30:00.287132 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxg9b\" (UniqueName: \"kubernetes.io/projected/5c0d4c12-ebd6-4020-aa27-d5dfda947b44-kube-api-access-bxg9b\") pod \"collect-profiles-29396490-2ksls\" (UID: \"5c0d4c12-ebd6-4020-aa27-d5dfda947b44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls" Nov 22 05:30:00 crc kubenswrapper[4818]: I1122 05:30:00.287222 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c0d4c12-ebd6-4020-aa27-d5dfda947b44-config-volume\") pod \"collect-profiles-29396490-2ksls\" (UID: \"5c0d4c12-ebd6-4020-aa27-d5dfda947b44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls" Nov 22 05:30:00 crc kubenswrapper[4818]: I1122 05:30:00.390915 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5c0d4c12-ebd6-4020-aa27-d5dfda947b44-secret-volume\") pod \"collect-profiles-29396490-2ksls\" (UID: \"5c0d4c12-ebd6-4020-aa27-d5dfda947b44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls" Nov 22 05:30:00 crc kubenswrapper[4818]: I1122 05:30:00.391016 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxg9b\" (UniqueName: \"kubernetes.io/projected/5c0d4c12-ebd6-4020-aa27-d5dfda947b44-kube-api-access-bxg9b\") pod \"collect-profiles-29396490-2ksls\" (UID: \"5c0d4c12-ebd6-4020-aa27-d5dfda947b44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls" Nov 22 05:30:00 crc kubenswrapper[4818]: I1122 05:30:00.391107 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c0d4c12-ebd6-4020-aa27-d5dfda947b44-config-volume\") pod \"collect-profiles-29396490-2ksls\" (UID: \"5c0d4c12-ebd6-4020-aa27-d5dfda947b44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls" Nov 22 05:30:00 crc kubenswrapper[4818]: I1122 05:30:00.391979 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c0d4c12-ebd6-4020-aa27-d5dfda947b44-config-volume\") pod \"collect-profiles-29396490-2ksls\" (UID: \"5c0d4c12-ebd6-4020-aa27-d5dfda947b44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls" Nov 22 05:30:00 crc kubenswrapper[4818]: I1122 05:30:00.399918 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5c0d4c12-ebd6-4020-aa27-d5dfda947b44-secret-volume\") pod \"collect-profiles-29396490-2ksls\" (UID: \"5c0d4c12-ebd6-4020-aa27-d5dfda947b44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls" Nov 22 05:30:00 crc kubenswrapper[4818]: I1122 05:30:00.408662 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxg9b\" (UniqueName: \"kubernetes.io/projected/5c0d4c12-ebd6-4020-aa27-d5dfda947b44-kube-api-access-bxg9b\") pod \"collect-profiles-29396490-2ksls\" (UID: \"5c0d4c12-ebd6-4020-aa27-d5dfda947b44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls" Nov 22 05:30:00 crc kubenswrapper[4818]: I1122 05:30:00.499101 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls" Nov 22 05:30:00 crc kubenswrapper[4818]: I1122 05:30:00.935643 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls"] Nov 22 05:30:00 crc kubenswrapper[4818]: W1122 05:30:00.942463 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c0d4c12_ebd6_4020_aa27_d5dfda947b44.slice/crio-c98dd69de7cb5a35b7da0628601f537bb268f5c906ead8691ee330fcfdac07ad WatchSource:0}: Error finding container c98dd69de7cb5a35b7da0628601f537bb268f5c906ead8691ee330fcfdac07ad: Status 404 returned error can't find the container with id c98dd69de7cb5a35b7da0628601f537bb268f5c906ead8691ee330fcfdac07ad Nov 22 05:30:01 crc kubenswrapper[4818]: I1122 05:30:01.083140 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls" event={"ID":"5c0d4c12-ebd6-4020-aa27-d5dfda947b44","Type":"ContainerStarted","Data":"c98dd69de7cb5a35b7da0628601f537bb268f5c906ead8691ee330fcfdac07ad"} Nov 22 05:30:02 crc kubenswrapper[4818]: I1122 05:30:02.092018 4818 generic.go:334] "Generic (PLEG): container finished" podID="5c0d4c12-ebd6-4020-aa27-d5dfda947b44" containerID="c414d1cd4492199df9c403e7ff11723287375fc6caa6b7abd4b200bdd18ecc2d" exitCode=0 Nov 22 05:30:02 crc kubenswrapper[4818]: I1122 05:30:02.092085 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls" event={"ID":"5c0d4c12-ebd6-4020-aa27-d5dfda947b44","Type":"ContainerDied","Data":"c414d1cd4492199df9c403e7ff11723287375fc6caa6b7abd4b200bdd18ecc2d"} Nov 22 05:30:03 crc kubenswrapper[4818]: I1122 05:30:03.292216 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:30:03 crc kubenswrapper[4818]: E1122 05:30:03.292751 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:30:03 crc kubenswrapper[4818]: I1122 05:30:03.520489 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls" Nov 22 05:30:03 crc kubenswrapper[4818]: I1122 05:30:03.549320 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxg9b\" (UniqueName: \"kubernetes.io/projected/5c0d4c12-ebd6-4020-aa27-d5dfda947b44-kube-api-access-bxg9b\") pod \"5c0d4c12-ebd6-4020-aa27-d5dfda947b44\" (UID: \"5c0d4c12-ebd6-4020-aa27-d5dfda947b44\") " Nov 22 05:30:03 crc kubenswrapper[4818]: I1122 05:30:03.549381 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5c0d4c12-ebd6-4020-aa27-d5dfda947b44-secret-volume\") pod \"5c0d4c12-ebd6-4020-aa27-d5dfda947b44\" (UID: \"5c0d4c12-ebd6-4020-aa27-d5dfda947b44\") " Nov 22 05:30:03 crc kubenswrapper[4818]: I1122 05:30:03.549469 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c0d4c12-ebd6-4020-aa27-d5dfda947b44-config-volume\") pod \"5c0d4c12-ebd6-4020-aa27-d5dfda947b44\" (UID: \"5c0d4c12-ebd6-4020-aa27-d5dfda947b44\") " Nov 22 05:30:03 crc kubenswrapper[4818]: I1122 05:30:03.550520 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c0d4c12-ebd6-4020-aa27-d5dfda947b44-config-volume" (OuterVolumeSpecName: "config-volume") pod "5c0d4c12-ebd6-4020-aa27-d5dfda947b44" (UID: "5c0d4c12-ebd6-4020-aa27-d5dfda947b44"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:30:03 crc kubenswrapper[4818]: I1122 05:30:03.551492 4818 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c0d4c12-ebd6-4020-aa27-d5dfda947b44-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 05:30:03 crc kubenswrapper[4818]: I1122 05:30:03.560404 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c0d4c12-ebd6-4020-aa27-d5dfda947b44-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5c0d4c12-ebd6-4020-aa27-d5dfda947b44" (UID: "5c0d4c12-ebd6-4020-aa27-d5dfda947b44"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:30:03 crc kubenswrapper[4818]: I1122 05:30:03.560461 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c0d4c12-ebd6-4020-aa27-d5dfda947b44-kube-api-access-bxg9b" (OuterVolumeSpecName: "kube-api-access-bxg9b") pod "5c0d4c12-ebd6-4020-aa27-d5dfda947b44" (UID: "5c0d4c12-ebd6-4020-aa27-d5dfda947b44"). InnerVolumeSpecName "kube-api-access-bxg9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:30:03 crc kubenswrapper[4818]: I1122 05:30:03.653534 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxg9b\" (UniqueName: \"kubernetes.io/projected/5c0d4c12-ebd6-4020-aa27-d5dfda947b44-kube-api-access-bxg9b\") on node \"crc\" DevicePath \"\"" Nov 22 05:30:03 crc kubenswrapper[4818]: I1122 05:30:03.653722 4818 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5c0d4c12-ebd6-4020-aa27-d5dfda947b44-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 05:30:04 crc kubenswrapper[4818]: I1122 05:30:04.106926 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls" event={"ID":"5c0d4c12-ebd6-4020-aa27-d5dfda947b44","Type":"ContainerDied","Data":"c98dd69de7cb5a35b7da0628601f537bb268f5c906ead8691ee330fcfdac07ad"} Nov 22 05:30:04 crc kubenswrapper[4818]: I1122 05:30:04.106973 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls" Nov 22 05:30:04 crc kubenswrapper[4818]: I1122 05:30:04.106975 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c98dd69de7cb5a35b7da0628601f537bb268f5c906ead8691ee330fcfdac07ad" Nov 22 05:30:04 crc kubenswrapper[4818]: I1122 05:30:04.612531 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4"] Nov 22 05:30:04 crc kubenswrapper[4818]: I1122 05:30:04.620681 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396445-v6dn4"] Nov 22 05:30:06 crc kubenswrapper[4818]: I1122 05:30:06.304846 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="751d9838-4c34-4251-b3a6-1d808f98a034" path="/var/lib/kubelet/pods/751d9838-4c34-4251-b3a6-1d808f98a034/volumes" Nov 22 05:30:07 crc kubenswrapper[4818]: I1122 05:30:07.142620 4818 generic.go:334] "Generic (PLEG): container finished" podID="597bc073-9fa3-485c-b23d-8f7b5ea21da9" containerID="74281e6854517e3d04a170a5ccbc81b5b2ac60259a5492e0a70a7c3f555a0fed" exitCode=0 Nov 22 05:30:07 crc kubenswrapper[4818]: I1122 05:30:07.142678 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" event={"ID":"597bc073-9fa3-485c-b23d-8f7b5ea21da9","Type":"ContainerDied","Data":"74281e6854517e3d04a170a5ccbc81b5b2ac60259a5492e0a70a7c3f555a0fed"} Nov 22 05:30:08 crc kubenswrapper[4818]: I1122 05:30:08.547755 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" Nov 22 05:30:08 crc kubenswrapper[4818]: I1122 05:30:08.738819 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-inventory\") pod \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\" (UID: \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\") " Nov 22 05:30:08 crc kubenswrapper[4818]: I1122 05:30:08.739000 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbx5r\" (UniqueName: \"kubernetes.io/projected/597bc073-9fa3-485c-b23d-8f7b5ea21da9-kube-api-access-vbx5r\") pod \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\" (UID: \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\") " Nov 22 05:30:08 crc kubenswrapper[4818]: I1122 05:30:08.739037 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-bootstrap-combined-ca-bundle\") pod \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\" (UID: \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\") " Nov 22 05:30:08 crc kubenswrapper[4818]: I1122 05:30:08.739059 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-ssh-key\") pod \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\" (UID: \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\") " Nov 22 05:30:08 crc kubenswrapper[4818]: I1122 05:30:08.739133 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-ceph\") pod \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\" (UID: \"597bc073-9fa3-485c-b23d-8f7b5ea21da9\") " Nov 22 05:30:08 crc kubenswrapper[4818]: I1122 05:30:08.745208 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/597bc073-9fa3-485c-b23d-8f7b5ea21da9-kube-api-access-vbx5r" (OuterVolumeSpecName: "kube-api-access-vbx5r") pod "597bc073-9fa3-485c-b23d-8f7b5ea21da9" (UID: "597bc073-9fa3-485c-b23d-8f7b5ea21da9"). InnerVolumeSpecName "kube-api-access-vbx5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:30:08 crc kubenswrapper[4818]: I1122 05:30:08.752469 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-ceph" (OuterVolumeSpecName: "ceph") pod "597bc073-9fa3-485c-b23d-8f7b5ea21da9" (UID: "597bc073-9fa3-485c-b23d-8f7b5ea21da9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:30:08 crc kubenswrapper[4818]: I1122 05:30:08.755454 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "597bc073-9fa3-485c-b23d-8f7b5ea21da9" (UID: "597bc073-9fa3-485c-b23d-8f7b5ea21da9"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:30:08 crc kubenswrapper[4818]: I1122 05:30:08.767506 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "597bc073-9fa3-485c-b23d-8f7b5ea21da9" (UID: "597bc073-9fa3-485c-b23d-8f7b5ea21da9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:30:08 crc kubenswrapper[4818]: I1122 05:30:08.799715 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-inventory" (OuterVolumeSpecName: "inventory") pod "597bc073-9fa3-485c-b23d-8f7b5ea21da9" (UID: "597bc073-9fa3-485c-b23d-8f7b5ea21da9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:30:08 crc kubenswrapper[4818]: I1122 05:30:08.843148 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:30:08 crc kubenswrapper[4818]: I1122 05:30:08.843181 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbx5r\" (UniqueName: \"kubernetes.io/projected/597bc073-9fa3-485c-b23d-8f7b5ea21da9-kube-api-access-vbx5r\") on node \"crc\" DevicePath \"\"" Nov 22 05:30:08 crc kubenswrapper[4818]: I1122 05:30:08.843193 4818 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:30:08 crc kubenswrapper[4818]: I1122 05:30:08.843203 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:30:08 crc kubenswrapper[4818]: I1122 05:30:08.843211 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/597bc073-9fa3-485c-b23d-8f7b5ea21da9-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.164069 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" event={"ID":"597bc073-9fa3-485c-b23d-8f7b5ea21da9","Type":"ContainerDied","Data":"dbcc23b10c35186df0b90c73eef5b12989d7529146a08ad260c20571c4aae0dc"} Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.164524 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbcc23b10c35186df0b90c73eef5b12989d7529146a08ad260c20571c4aae0dc" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.164158 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.270563 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6"] Nov 22 05:30:09 crc kubenswrapper[4818]: E1122 05:30:09.270989 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="597bc073-9fa3-485c-b23d-8f7b5ea21da9" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.271013 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="597bc073-9fa3-485c-b23d-8f7b5ea21da9" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 22 05:30:09 crc kubenswrapper[4818]: E1122 05:30:09.271042 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c0d4c12-ebd6-4020-aa27-d5dfda947b44" containerName="collect-profiles" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.271054 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c0d4c12-ebd6-4020-aa27-d5dfda947b44" containerName="collect-profiles" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.271310 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="597bc073-9fa3-485c-b23d-8f7b5ea21da9" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.271336 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c0d4c12-ebd6-4020-aa27-d5dfda947b44" containerName="collect-profiles" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.272055 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.275073 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.275168 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.275781 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.275844 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.275864 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.299113 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6"] Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.454775 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e6cd425-5f7c-4f17-864b-2e3d62851c58-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6\" (UID: \"4e6cd425-5f7c-4f17-864b-2e3d62851c58\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.454964 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e6cd425-5f7c-4f17-864b-2e3d62851c58-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6\" (UID: \"4e6cd425-5f7c-4f17-864b-2e3d62851c58\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.455429 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4e6cd425-5f7c-4f17-864b-2e3d62851c58-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6\" (UID: \"4e6cd425-5f7c-4f17-864b-2e3d62851c58\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.455730 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxgln\" (UniqueName: \"kubernetes.io/projected/4e6cd425-5f7c-4f17-864b-2e3d62851c58-kube-api-access-zxgln\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6\" (UID: \"4e6cd425-5f7c-4f17-864b-2e3d62851c58\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.557857 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4e6cd425-5f7c-4f17-864b-2e3d62851c58-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6\" (UID: \"4e6cd425-5f7c-4f17-864b-2e3d62851c58\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.557958 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxgln\" (UniqueName: \"kubernetes.io/projected/4e6cd425-5f7c-4f17-864b-2e3d62851c58-kube-api-access-zxgln\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6\" (UID: \"4e6cd425-5f7c-4f17-864b-2e3d62851c58\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.558056 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e6cd425-5f7c-4f17-864b-2e3d62851c58-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6\" (UID: \"4e6cd425-5f7c-4f17-864b-2e3d62851c58\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.558100 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e6cd425-5f7c-4f17-864b-2e3d62851c58-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6\" (UID: \"4e6cd425-5f7c-4f17-864b-2e3d62851c58\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.564245 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e6cd425-5f7c-4f17-864b-2e3d62851c58-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6\" (UID: \"4e6cd425-5f7c-4f17-864b-2e3d62851c58\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.564539 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4e6cd425-5f7c-4f17-864b-2e3d62851c58-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6\" (UID: \"4e6cd425-5f7c-4f17-864b-2e3d62851c58\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.568243 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e6cd425-5f7c-4f17-864b-2e3d62851c58-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6\" (UID: \"4e6cd425-5f7c-4f17-864b-2e3d62851c58\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.592869 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxgln\" (UniqueName: \"kubernetes.io/projected/4e6cd425-5f7c-4f17-864b-2e3d62851c58-kube-api-access-zxgln\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6\" (UID: \"4e6cd425-5f7c-4f17-864b-2e3d62851c58\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.594614 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" Nov 22 05:30:09 crc kubenswrapper[4818]: I1122 05:30:09.994277 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6"] Nov 22 05:30:10 crc kubenswrapper[4818]: I1122 05:30:10.176314 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" event={"ID":"4e6cd425-5f7c-4f17-864b-2e3d62851c58","Type":"ContainerStarted","Data":"285d4669021793ba571631bb3e3551e057136e9b94e0d6d8967be5f329eee13d"} Nov 22 05:30:11 crc kubenswrapper[4818]: I1122 05:30:11.190688 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" event={"ID":"4e6cd425-5f7c-4f17-864b-2e3d62851c58","Type":"ContainerStarted","Data":"d9e46524146f94705a87619e9afabfee337c5cc117ba946b860a85f09eda8171"} Nov 22 05:30:11 crc kubenswrapper[4818]: I1122 05:30:11.212584 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" podStartSLOduration=1.7387729859999999 podStartE2EDuration="2.212565293s" podCreationTimestamp="2025-11-22 05:30:09 +0000 UTC" firstStartedPulling="2025-11-22 05:30:10.007760494 +0000 UTC m=+2562.582177031" lastFinishedPulling="2025-11-22 05:30:10.481552811 +0000 UTC m=+2563.055969338" observedRunningTime="2025-11-22 05:30:11.207695893 +0000 UTC m=+2563.782112430" watchObservedRunningTime="2025-11-22 05:30:11.212565293 +0000 UTC m=+2563.786981830" Nov 22 05:30:18 crc kubenswrapper[4818]: I1122 05:30:18.297114 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:30:18 crc kubenswrapper[4818]: E1122 05:30:18.297930 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:30:32 crc kubenswrapper[4818]: I1122 05:30:32.294686 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:30:32 crc kubenswrapper[4818]: E1122 05:30:32.295567 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:30:38 crc kubenswrapper[4818]: I1122 05:30:38.473086 4818 generic.go:334] "Generic (PLEG): container finished" podID="4e6cd425-5f7c-4f17-864b-2e3d62851c58" containerID="d9e46524146f94705a87619e9afabfee337c5cc117ba946b860a85f09eda8171" exitCode=0 Nov 22 05:30:38 crc kubenswrapper[4818]: I1122 05:30:38.473182 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" event={"ID":"4e6cd425-5f7c-4f17-864b-2e3d62851c58","Type":"ContainerDied","Data":"d9e46524146f94705a87619e9afabfee337c5cc117ba946b860a85f09eda8171"} Nov 22 05:30:39 crc kubenswrapper[4818]: I1122 05:30:39.925744 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.018065 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e6cd425-5f7c-4f17-864b-2e3d62851c58-inventory\") pod \"4e6cd425-5f7c-4f17-864b-2e3d62851c58\" (UID: \"4e6cd425-5f7c-4f17-864b-2e3d62851c58\") " Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.018120 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4e6cd425-5f7c-4f17-864b-2e3d62851c58-ceph\") pod \"4e6cd425-5f7c-4f17-864b-2e3d62851c58\" (UID: \"4e6cd425-5f7c-4f17-864b-2e3d62851c58\") " Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.018175 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxgln\" (UniqueName: \"kubernetes.io/projected/4e6cd425-5f7c-4f17-864b-2e3d62851c58-kube-api-access-zxgln\") pod \"4e6cd425-5f7c-4f17-864b-2e3d62851c58\" (UID: \"4e6cd425-5f7c-4f17-864b-2e3d62851c58\") " Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.018214 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e6cd425-5f7c-4f17-864b-2e3d62851c58-ssh-key\") pod \"4e6cd425-5f7c-4f17-864b-2e3d62851c58\" (UID: \"4e6cd425-5f7c-4f17-864b-2e3d62851c58\") " Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.024357 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e6cd425-5f7c-4f17-864b-2e3d62851c58-ceph" (OuterVolumeSpecName: "ceph") pod "4e6cd425-5f7c-4f17-864b-2e3d62851c58" (UID: "4e6cd425-5f7c-4f17-864b-2e3d62851c58"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.025559 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e6cd425-5f7c-4f17-864b-2e3d62851c58-kube-api-access-zxgln" (OuterVolumeSpecName: "kube-api-access-zxgln") pod "4e6cd425-5f7c-4f17-864b-2e3d62851c58" (UID: "4e6cd425-5f7c-4f17-864b-2e3d62851c58"). InnerVolumeSpecName "kube-api-access-zxgln". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.050356 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e6cd425-5f7c-4f17-864b-2e3d62851c58-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4e6cd425-5f7c-4f17-864b-2e3d62851c58" (UID: "4e6cd425-5f7c-4f17-864b-2e3d62851c58"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.071943 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e6cd425-5f7c-4f17-864b-2e3d62851c58-inventory" (OuterVolumeSpecName: "inventory") pod "4e6cd425-5f7c-4f17-864b-2e3d62851c58" (UID: "4e6cd425-5f7c-4f17-864b-2e3d62851c58"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.120215 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxgln\" (UniqueName: \"kubernetes.io/projected/4e6cd425-5f7c-4f17-864b-2e3d62851c58-kube-api-access-zxgln\") on node \"crc\" DevicePath \"\"" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.120261 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e6cd425-5f7c-4f17-864b-2e3d62851c58-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.120270 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e6cd425-5f7c-4f17-864b-2e3d62851c58-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.120280 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4e6cd425-5f7c-4f17-864b-2e3d62851c58-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.503134 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" event={"ID":"4e6cd425-5f7c-4f17-864b-2e3d62851c58","Type":"ContainerDied","Data":"285d4669021793ba571631bb3e3551e057136e9b94e0d6d8967be5f329eee13d"} Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.503222 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="285d4669021793ba571631bb3e3551e057136e9b94e0d6d8967be5f329eee13d" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.503245 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.595208 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx"] Nov 22 05:30:40 crc kubenswrapper[4818]: E1122 05:30:40.595848 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e6cd425-5f7c-4f17-864b-2e3d62851c58" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.595934 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e6cd425-5f7c-4f17-864b-2e3d62851c58" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.596977 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e6cd425-5f7c-4f17-864b-2e3d62851c58" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.605066 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.606936 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.607837 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.608452 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.608672 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.609342 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.620532 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx"] Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.658439 4818 scope.go:117] "RemoveContainer" containerID="b7ccf6acf2f43368232355bb645f785a164ce14e18e91b3babade5df387c1565" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.731611 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx\" (UID: \"68cb051d-9556-47e4-acc2-3bdb9b60ebdd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.731733 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cgmn\" (UniqueName: \"kubernetes.io/projected/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-kube-api-access-7cgmn\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx\" (UID: \"68cb051d-9556-47e4-acc2-3bdb9b60ebdd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.731967 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx\" (UID: \"68cb051d-9556-47e4-acc2-3bdb9b60ebdd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.732082 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx\" (UID: \"68cb051d-9556-47e4-acc2-3bdb9b60ebdd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.834132 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx\" (UID: \"68cb051d-9556-47e4-acc2-3bdb9b60ebdd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.834450 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx\" (UID: \"68cb051d-9556-47e4-acc2-3bdb9b60ebdd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.834583 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cgmn\" (UniqueName: \"kubernetes.io/projected/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-kube-api-access-7cgmn\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx\" (UID: \"68cb051d-9556-47e4-acc2-3bdb9b60ebdd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.834747 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx\" (UID: \"68cb051d-9556-47e4-acc2-3bdb9b60ebdd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.838629 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx\" (UID: \"68cb051d-9556-47e4-acc2-3bdb9b60ebdd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.840460 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx\" (UID: \"68cb051d-9556-47e4-acc2-3bdb9b60ebdd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.840667 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx\" (UID: \"68cb051d-9556-47e4-acc2-3bdb9b60ebdd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.873671 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cgmn\" (UniqueName: \"kubernetes.io/projected/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-kube-api-access-7cgmn\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx\" (UID: \"68cb051d-9556-47e4-acc2-3bdb9b60ebdd\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" Nov 22 05:30:40 crc kubenswrapper[4818]: I1122 05:30:40.942103 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" Nov 22 05:30:41 crc kubenswrapper[4818]: I1122 05:30:41.528064 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx"] Nov 22 05:30:42 crc kubenswrapper[4818]: I1122 05:30:42.528093 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" event={"ID":"68cb051d-9556-47e4-acc2-3bdb9b60ebdd","Type":"ContainerStarted","Data":"c2a5bb6937e0f67c7bf8d2d1b0209df9c4d4a9fb75123c10e2f77ccc45871ff3"} Nov 22 05:30:42 crc kubenswrapper[4818]: I1122 05:30:42.529138 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" event={"ID":"68cb051d-9556-47e4-acc2-3bdb9b60ebdd","Type":"ContainerStarted","Data":"773bce0248421edda3c42c84144a76591fe87bac3e59fe69467c1c29cbac96f0"} Nov 22 05:30:42 crc kubenswrapper[4818]: I1122 05:30:42.556380 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" podStartSLOduration=2.145504988 podStartE2EDuration="2.556346059s" podCreationTimestamp="2025-11-22 05:30:40 +0000 UTC" firstStartedPulling="2025-11-22 05:30:41.538034921 +0000 UTC m=+2594.112451458" lastFinishedPulling="2025-11-22 05:30:41.948875972 +0000 UTC m=+2594.523292529" observedRunningTime="2025-11-22 05:30:42.544192632 +0000 UTC m=+2595.118609159" watchObservedRunningTime="2025-11-22 05:30:42.556346059 +0000 UTC m=+2595.130762656" Nov 22 05:30:47 crc kubenswrapper[4818]: I1122 05:30:47.290621 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:30:47 crc kubenswrapper[4818]: E1122 05:30:47.291388 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:30:47 crc kubenswrapper[4818]: I1122 05:30:47.580907 4818 generic.go:334] "Generic (PLEG): container finished" podID="68cb051d-9556-47e4-acc2-3bdb9b60ebdd" containerID="c2a5bb6937e0f67c7bf8d2d1b0209df9c4d4a9fb75123c10e2f77ccc45871ff3" exitCode=0 Nov 22 05:30:47 crc kubenswrapper[4818]: I1122 05:30:47.580978 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" event={"ID":"68cb051d-9556-47e4-acc2-3bdb9b60ebdd","Type":"ContainerDied","Data":"c2a5bb6937e0f67c7bf8d2d1b0209df9c4d4a9fb75123c10e2f77ccc45871ff3"} Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.074391 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.203049 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-ceph\") pod \"68cb051d-9556-47e4-acc2-3bdb9b60ebdd\" (UID: \"68cb051d-9556-47e4-acc2-3bdb9b60ebdd\") " Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.203179 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-inventory\") pod \"68cb051d-9556-47e4-acc2-3bdb9b60ebdd\" (UID: \"68cb051d-9556-47e4-acc2-3bdb9b60ebdd\") " Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.203244 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-ssh-key\") pod \"68cb051d-9556-47e4-acc2-3bdb9b60ebdd\" (UID: \"68cb051d-9556-47e4-acc2-3bdb9b60ebdd\") " Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.203327 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cgmn\" (UniqueName: \"kubernetes.io/projected/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-kube-api-access-7cgmn\") pod \"68cb051d-9556-47e4-acc2-3bdb9b60ebdd\" (UID: \"68cb051d-9556-47e4-acc2-3bdb9b60ebdd\") " Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.208556 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-kube-api-access-7cgmn" (OuterVolumeSpecName: "kube-api-access-7cgmn") pod "68cb051d-9556-47e4-acc2-3bdb9b60ebdd" (UID: "68cb051d-9556-47e4-acc2-3bdb9b60ebdd"). InnerVolumeSpecName "kube-api-access-7cgmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.209924 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-ceph" (OuterVolumeSpecName: "ceph") pod "68cb051d-9556-47e4-acc2-3bdb9b60ebdd" (UID: "68cb051d-9556-47e4-acc2-3bdb9b60ebdd"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.235363 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-inventory" (OuterVolumeSpecName: "inventory") pod "68cb051d-9556-47e4-acc2-3bdb9b60ebdd" (UID: "68cb051d-9556-47e4-acc2-3bdb9b60ebdd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.237153 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "68cb051d-9556-47e4-acc2-3bdb9b60ebdd" (UID: "68cb051d-9556-47e4-acc2-3bdb9b60ebdd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.305775 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.305926 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.306001 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cgmn\" (UniqueName: \"kubernetes.io/projected/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-kube-api-access-7cgmn\") on node \"crc\" DevicePath \"\"" Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.306076 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/68cb051d-9556-47e4-acc2-3bdb9b60ebdd-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.906309 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" event={"ID":"68cb051d-9556-47e4-acc2-3bdb9b60ebdd","Type":"ContainerDied","Data":"773bce0248421edda3c42c84144a76591fe87bac3e59fe69467c1c29cbac96f0"} Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.906358 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="773bce0248421edda3c42c84144a76591fe87bac3e59fe69467c1c29cbac96f0" Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.906431 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx" Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.908511 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g"] Nov 22 05:30:49 crc kubenswrapper[4818]: E1122 05:30:49.909178 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68cb051d-9556-47e4-acc2-3bdb9b60ebdd" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.909212 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="68cb051d-9556-47e4-acc2-3bdb9b60ebdd" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.909590 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="68cb051d-9556-47e4-acc2-3bdb9b60ebdd" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.910935 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.913041 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.913492 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.913601 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g"] Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.914044 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.914375 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 05:30:49 crc kubenswrapper[4818]: I1122 05:30:49.915328 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:30:50 crc kubenswrapper[4818]: I1122 05:30:50.092298 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6pgv\" (UniqueName: \"kubernetes.io/projected/969421c2-6440-41e6-9866-c624dfe1cbae-kube-api-access-n6pgv\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fld7g\" (UID: \"969421c2-6440-41e6-9866-c624dfe1cbae\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" Nov 22 05:30:50 crc kubenswrapper[4818]: I1122 05:30:50.092810 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/969421c2-6440-41e6-9866-c624dfe1cbae-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fld7g\" (UID: \"969421c2-6440-41e6-9866-c624dfe1cbae\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" Nov 22 05:30:50 crc kubenswrapper[4818]: I1122 05:30:50.093020 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/969421c2-6440-41e6-9866-c624dfe1cbae-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fld7g\" (UID: \"969421c2-6440-41e6-9866-c624dfe1cbae\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" Nov 22 05:30:50 crc kubenswrapper[4818]: I1122 05:30:50.093070 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/969421c2-6440-41e6-9866-c624dfe1cbae-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fld7g\" (UID: \"969421c2-6440-41e6-9866-c624dfe1cbae\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" Nov 22 05:30:50 crc kubenswrapper[4818]: I1122 05:30:50.195452 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6pgv\" (UniqueName: \"kubernetes.io/projected/969421c2-6440-41e6-9866-c624dfe1cbae-kube-api-access-n6pgv\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fld7g\" (UID: \"969421c2-6440-41e6-9866-c624dfe1cbae\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" Nov 22 05:30:50 crc kubenswrapper[4818]: I1122 05:30:50.195961 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/969421c2-6440-41e6-9866-c624dfe1cbae-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fld7g\" (UID: \"969421c2-6440-41e6-9866-c624dfe1cbae\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" Nov 22 05:30:50 crc kubenswrapper[4818]: I1122 05:30:50.196409 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/969421c2-6440-41e6-9866-c624dfe1cbae-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fld7g\" (UID: \"969421c2-6440-41e6-9866-c624dfe1cbae\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" Nov 22 05:30:50 crc kubenswrapper[4818]: I1122 05:30:50.196466 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/969421c2-6440-41e6-9866-c624dfe1cbae-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fld7g\" (UID: \"969421c2-6440-41e6-9866-c624dfe1cbae\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" Nov 22 05:30:50 crc kubenswrapper[4818]: I1122 05:30:50.205939 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/969421c2-6440-41e6-9866-c624dfe1cbae-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fld7g\" (UID: \"969421c2-6440-41e6-9866-c624dfe1cbae\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" Nov 22 05:30:50 crc kubenswrapper[4818]: I1122 05:30:50.206444 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/969421c2-6440-41e6-9866-c624dfe1cbae-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fld7g\" (UID: \"969421c2-6440-41e6-9866-c624dfe1cbae\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" Nov 22 05:30:50 crc kubenswrapper[4818]: I1122 05:30:50.207749 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/969421c2-6440-41e6-9866-c624dfe1cbae-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fld7g\" (UID: \"969421c2-6440-41e6-9866-c624dfe1cbae\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" Nov 22 05:30:50 crc kubenswrapper[4818]: I1122 05:30:50.224227 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6pgv\" (UniqueName: \"kubernetes.io/projected/969421c2-6440-41e6-9866-c624dfe1cbae-kube-api-access-n6pgv\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fld7g\" (UID: \"969421c2-6440-41e6-9866-c624dfe1cbae\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" Nov 22 05:30:50 crc kubenswrapper[4818]: I1122 05:30:50.234302 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" Nov 22 05:30:50 crc kubenswrapper[4818]: I1122 05:30:50.579653 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g"] Nov 22 05:30:50 crc kubenswrapper[4818]: I1122 05:30:50.584096 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 05:30:50 crc kubenswrapper[4818]: I1122 05:30:50.919936 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" event={"ID":"969421c2-6440-41e6-9866-c624dfe1cbae","Type":"ContainerStarted","Data":"af697b13213fad4e892793a535fc347ebedf39a7da099d54af5b7cf9069667ae"} Nov 22 05:30:51 crc kubenswrapper[4818]: I1122 05:30:51.930727 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" event={"ID":"969421c2-6440-41e6-9866-c624dfe1cbae","Type":"ContainerStarted","Data":"b43ef81ad49a7bcfee66d968929d9b6eb3cafce1ff330c451d67c3b71bcb8dd6"} Nov 22 05:30:51 crc kubenswrapper[4818]: I1122 05:30:51.946788 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" podStartSLOduration=2.51231706 podStartE2EDuration="2.946763907s" podCreationTimestamp="2025-11-22 05:30:49 +0000 UTC" firstStartedPulling="2025-11-22 05:30:50.583897612 +0000 UTC m=+2603.158314139" lastFinishedPulling="2025-11-22 05:30:51.018344459 +0000 UTC m=+2603.592760986" observedRunningTime="2025-11-22 05:30:51.945509533 +0000 UTC m=+2604.519926070" watchObservedRunningTime="2025-11-22 05:30:51.946763907 +0000 UTC m=+2604.521180444" Nov 22 05:31:00 crc kubenswrapper[4818]: I1122 05:31:00.291614 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:31:00 crc kubenswrapper[4818]: E1122 05:31:00.292627 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:31:12 crc kubenswrapper[4818]: I1122 05:31:12.291716 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:31:12 crc kubenswrapper[4818]: E1122 05:31:12.293163 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:31:26 crc kubenswrapper[4818]: I1122 05:31:26.291390 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:31:26 crc kubenswrapper[4818]: E1122 05:31:26.292351 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:31:32 crc kubenswrapper[4818]: I1122 05:31:32.393353 4818 generic.go:334] "Generic (PLEG): container finished" podID="969421c2-6440-41e6-9866-c624dfe1cbae" containerID="b43ef81ad49a7bcfee66d968929d9b6eb3cafce1ff330c451d67c3b71bcb8dd6" exitCode=0 Nov 22 05:31:32 crc kubenswrapper[4818]: I1122 05:31:32.393486 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" event={"ID":"969421c2-6440-41e6-9866-c624dfe1cbae","Type":"ContainerDied","Data":"b43ef81ad49a7bcfee66d968929d9b6eb3cafce1ff330c451d67c3b71bcb8dd6"} Nov 22 05:31:33 crc kubenswrapper[4818]: I1122 05:31:33.852135 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" Nov 22 05:31:33 crc kubenswrapper[4818]: I1122 05:31:33.998513 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/969421c2-6440-41e6-9866-c624dfe1cbae-inventory\") pod \"969421c2-6440-41e6-9866-c624dfe1cbae\" (UID: \"969421c2-6440-41e6-9866-c624dfe1cbae\") " Nov 22 05:31:33 crc kubenswrapper[4818]: I1122 05:31:33.998683 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/969421c2-6440-41e6-9866-c624dfe1cbae-ceph\") pod \"969421c2-6440-41e6-9866-c624dfe1cbae\" (UID: \"969421c2-6440-41e6-9866-c624dfe1cbae\") " Nov 22 05:31:33 crc kubenswrapper[4818]: I1122 05:31:33.998755 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/969421c2-6440-41e6-9866-c624dfe1cbae-ssh-key\") pod \"969421c2-6440-41e6-9866-c624dfe1cbae\" (UID: \"969421c2-6440-41e6-9866-c624dfe1cbae\") " Nov 22 05:31:33 crc kubenswrapper[4818]: I1122 05:31:33.998824 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6pgv\" (UniqueName: \"kubernetes.io/projected/969421c2-6440-41e6-9866-c624dfe1cbae-kube-api-access-n6pgv\") pod \"969421c2-6440-41e6-9866-c624dfe1cbae\" (UID: \"969421c2-6440-41e6-9866-c624dfe1cbae\") " Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.006678 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/969421c2-6440-41e6-9866-c624dfe1cbae-ceph" (OuterVolumeSpecName: "ceph") pod "969421c2-6440-41e6-9866-c624dfe1cbae" (UID: "969421c2-6440-41e6-9866-c624dfe1cbae"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.007609 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/969421c2-6440-41e6-9866-c624dfe1cbae-kube-api-access-n6pgv" (OuterVolumeSpecName: "kube-api-access-n6pgv") pod "969421c2-6440-41e6-9866-c624dfe1cbae" (UID: "969421c2-6440-41e6-9866-c624dfe1cbae"). InnerVolumeSpecName "kube-api-access-n6pgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.028535 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/969421c2-6440-41e6-9866-c624dfe1cbae-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "969421c2-6440-41e6-9866-c624dfe1cbae" (UID: "969421c2-6440-41e6-9866-c624dfe1cbae"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.029665 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/969421c2-6440-41e6-9866-c624dfe1cbae-inventory" (OuterVolumeSpecName: "inventory") pod "969421c2-6440-41e6-9866-c624dfe1cbae" (UID: "969421c2-6440-41e6-9866-c624dfe1cbae"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.100356 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6pgv\" (UniqueName: \"kubernetes.io/projected/969421c2-6440-41e6-9866-c624dfe1cbae-kube-api-access-n6pgv\") on node \"crc\" DevicePath \"\"" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.100525 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/969421c2-6440-41e6-9866-c624dfe1cbae-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.100605 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/969421c2-6440-41e6-9866-c624dfe1cbae-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.100659 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/969421c2-6440-41e6-9866-c624dfe1cbae-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.412915 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" event={"ID":"969421c2-6440-41e6-9866-c624dfe1cbae","Type":"ContainerDied","Data":"af697b13213fad4e892793a535fc347ebedf39a7da099d54af5b7cf9069667ae"} Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.412952 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af697b13213fad4e892793a535fc347ebedf39a7da099d54af5b7cf9069667ae" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.412967 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fld7g" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.503226 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8"] Nov 22 05:31:34 crc kubenswrapper[4818]: E1122 05:31:34.503583 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="969421c2-6440-41e6-9866-c624dfe1cbae" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.503600 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="969421c2-6440-41e6-9866-c624dfe1cbae" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.503797 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="969421c2-6440-41e6-9866-c624dfe1cbae" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.504353 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.510432 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.510533 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.511033 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.511585 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.512898 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.524428 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8"] Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.611055 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fe8b800-5e58-48ae-8da3-eae014028909-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8\" (UID: \"5fe8b800-5e58-48ae-8da3-eae014028909\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.611137 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5fe8b800-5e58-48ae-8da3-eae014028909-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8\" (UID: \"5fe8b800-5e58-48ae-8da3-eae014028909\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.611187 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fe8b800-5e58-48ae-8da3-eae014028909-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8\" (UID: \"5fe8b800-5e58-48ae-8da3-eae014028909\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.611242 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9b6z\" (UniqueName: \"kubernetes.io/projected/5fe8b800-5e58-48ae-8da3-eae014028909-kube-api-access-r9b6z\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8\" (UID: \"5fe8b800-5e58-48ae-8da3-eae014028909\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.713206 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9b6z\" (UniqueName: \"kubernetes.io/projected/5fe8b800-5e58-48ae-8da3-eae014028909-kube-api-access-r9b6z\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8\" (UID: \"5fe8b800-5e58-48ae-8da3-eae014028909\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.713631 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fe8b800-5e58-48ae-8da3-eae014028909-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8\" (UID: \"5fe8b800-5e58-48ae-8da3-eae014028909\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.713809 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5fe8b800-5e58-48ae-8da3-eae014028909-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8\" (UID: \"5fe8b800-5e58-48ae-8da3-eae014028909\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.713985 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fe8b800-5e58-48ae-8da3-eae014028909-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8\" (UID: \"5fe8b800-5e58-48ae-8da3-eae014028909\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.716901 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5fe8b800-5e58-48ae-8da3-eae014028909-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8\" (UID: \"5fe8b800-5e58-48ae-8da3-eae014028909\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.720812 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fe8b800-5e58-48ae-8da3-eae014028909-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8\" (UID: \"5fe8b800-5e58-48ae-8da3-eae014028909\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.721005 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fe8b800-5e58-48ae-8da3-eae014028909-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8\" (UID: \"5fe8b800-5e58-48ae-8da3-eae014028909\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.732404 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9b6z\" (UniqueName: \"kubernetes.io/projected/5fe8b800-5e58-48ae-8da3-eae014028909-kube-api-access-r9b6z\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8\" (UID: \"5fe8b800-5e58-48ae-8da3-eae014028909\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" Nov 22 05:31:34 crc kubenswrapper[4818]: I1122 05:31:34.818541 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" Nov 22 05:31:35 crc kubenswrapper[4818]: I1122 05:31:35.402728 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8"] Nov 22 05:31:35 crc kubenswrapper[4818]: I1122 05:31:35.435902 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" event={"ID":"5fe8b800-5e58-48ae-8da3-eae014028909","Type":"ContainerStarted","Data":"809d98115cc89d27eed04f2fd13a23affd4656fa7de35743694fcbf4c993e39b"} Nov 22 05:31:36 crc kubenswrapper[4818]: I1122 05:31:36.445633 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" event={"ID":"5fe8b800-5e58-48ae-8da3-eae014028909","Type":"ContainerStarted","Data":"d045b77caac833645af7bf9c69c184cfc3e8f0b6cd2ee5e0c20e019feb33730c"} Nov 22 05:31:36 crc kubenswrapper[4818]: I1122 05:31:36.470468 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" podStartSLOduration=1.957320709 podStartE2EDuration="2.470451134s" podCreationTimestamp="2025-11-22 05:31:34 +0000 UTC" firstStartedPulling="2025-11-22 05:31:35.423426194 +0000 UTC m=+2647.997842741" lastFinishedPulling="2025-11-22 05:31:35.936556639 +0000 UTC m=+2648.510973166" observedRunningTime="2025-11-22 05:31:36.469943571 +0000 UTC m=+2649.044360098" watchObservedRunningTime="2025-11-22 05:31:36.470451134 +0000 UTC m=+2649.044867661" Nov 22 05:31:38 crc kubenswrapper[4818]: I1122 05:31:38.296953 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:31:38 crc kubenswrapper[4818]: E1122 05:31:38.297744 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:31:40 crc kubenswrapper[4818]: I1122 05:31:40.486076 4818 generic.go:334] "Generic (PLEG): container finished" podID="5fe8b800-5e58-48ae-8da3-eae014028909" containerID="d045b77caac833645af7bf9c69c184cfc3e8f0b6cd2ee5e0c20e019feb33730c" exitCode=0 Nov 22 05:31:40 crc kubenswrapper[4818]: I1122 05:31:40.486135 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" event={"ID":"5fe8b800-5e58-48ae-8da3-eae014028909","Type":"ContainerDied","Data":"d045b77caac833645af7bf9c69c184cfc3e8f0b6cd2ee5e0c20e019feb33730c"} Nov 22 05:31:41 crc kubenswrapper[4818]: I1122 05:31:41.929002 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.096474 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fe8b800-5e58-48ae-8da3-eae014028909-ssh-key\") pod \"5fe8b800-5e58-48ae-8da3-eae014028909\" (UID: \"5fe8b800-5e58-48ae-8da3-eae014028909\") " Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.096563 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5fe8b800-5e58-48ae-8da3-eae014028909-ceph\") pod \"5fe8b800-5e58-48ae-8da3-eae014028909\" (UID: \"5fe8b800-5e58-48ae-8da3-eae014028909\") " Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.096642 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9b6z\" (UniqueName: \"kubernetes.io/projected/5fe8b800-5e58-48ae-8da3-eae014028909-kube-api-access-r9b6z\") pod \"5fe8b800-5e58-48ae-8da3-eae014028909\" (UID: \"5fe8b800-5e58-48ae-8da3-eae014028909\") " Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.096728 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fe8b800-5e58-48ae-8da3-eae014028909-inventory\") pod \"5fe8b800-5e58-48ae-8da3-eae014028909\" (UID: \"5fe8b800-5e58-48ae-8da3-eae014028909\") " Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.102240 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe8b800-5e58-48ae-8da3-eae014028909-kube-api-access-r9b6z" (OuterVolumeSpecName: "kube-api-access-r9b6z") pod "5fe8b800-5e58-48ae-8da3-eae014028909" (UID: "5fe8b800-5e58-48ae-8da3-eae014028909"). InnerVolumeSpecName "kube-api-access-r9b6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.102945 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe8b800-5e58-48ae-8da3-eae014028909-ceph" (OuterVolumeSpecName: "ceph") pod "5fe8b800-5e58-48ae-8da3-eae014028909" (UID: "5fe8b800-5e58-48ae-8da3-eae014028909"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.131680 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe8b800-5e58-48ae-8da3-eae014028909-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5fe8b800-5e58-48ae-8da3-eae014028909" (UID: "5fe8b800-5e58-48ae-8da3-eae014028909"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.155084 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe8b800-5e58-48ae-8da3-eae014028909-inventory" (OuterVolumeSpecName: "inventory") pod "5fe8b800-5e58-48ae-8da3-eae014028909" (UID: "5fe8b800-5e58-48ae-8da3-eae014028909"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.199567 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fe8b800-5e58-48ae-8da3-eae014028909-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.199618 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5fe8b800-5e58-48ae-8da3-eae014028909-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.199637 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9b6z\" (UniqueName: \"kubernetes.io/projected/5fe8b800-5e58-48ae-8da3-eae014028909-kube-api-access-r9b6z\") on node \"crc\" DevicePath \"\"" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.199656 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fe8b800-5e58-48ae-8da3-eae014028909-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.508140 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" event={"ID":"5fe8b800-5e58-48ae-8da3-eae014028909","Type":"ContainerDied","Data":"809d98115cc89d27eed04f2fd13a23affd4656fa7de35743694fcbf4c993e39b"} Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.508198 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="809d98115cc89d27eed04f2fd13a23affd4656fa7de35743694fcbf4c993e39b" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.508235 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.596109 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z"] Nov 22 05:31:42 crc kubenswrapper[4818]: E1122 05:31:42.596551 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe8b800-5e58-48ae-8da3-eae014028909" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.596572 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe8b800-5e58-48ae-8da3-eae014028909" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.596838 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fe8b800-5e58-48ae-8da3-eae014028909" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.597550 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.600378 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.600491 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.601510 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.604062 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.615217 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.615520 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z"] Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.708488 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3f48375-4d0b-4876-b4e8-64a9fa970b63-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z\" (UID: \"d3f48375-4d0b-4876-b4e8-64a9fa970b63\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.708573 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5wdk\" (UniqueName: \"kubernetes.io/projected/d3f48375-4d0b-4876-b4e8-64a9fa970b63-kube-api-access-g5wdk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z\" (UID: \"d3f48375-4d0b-4876-b4e8-64a9fa970b63\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.709008 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d3f48375-4d0b-4876-b4e8-64a9fa970b63-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z\" (UID: \"d3f48375-4d0b-4876-b4e8-64a9fa970b63\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.709084 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3f48375-4d0b-4876-b4e8-64a9fa970b63-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z\" (UID: \"d3f48375-4d0b-4876-b4e8-64a9fa970b63\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.811128 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d3f48375-4d0b-4876-b4e8-64a9fa970b63-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z\" (UID: \"d3f48375-4d0b-4876-b4e8-64a9fa970b63\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.812448 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3f48375-4d0b-4876-b4e8-64a9fa970b63-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z\" (UID: \"d3f48375-4d0b-4876-b4e8-64a9fa970b63\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.812685 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3f48375-4d0b-4876-b4e8-64a9fa970b63-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z\" (UID: \"d3f48375-4d0b-4876-b4e8-64a9fa970b63\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.812747 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5wdk\" (UniqueName: \"kubernetes.io/projected/d3f48375-4d0b-4876-b4e8-64a9fa970b63-kube-api-access-g5wdk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z\" (UID: \"d3f48375-4d0b-4876-b4e8-64a9fa970b63\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.819362 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3f48375-4d0b-4876-b4e8-64a9fa970b63-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z\" (UID: \"d3f48375-4d0b-4876-b4e8-64a9fa970b63\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.821588 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d3f48375-4d0b-4876-b4e8-64a9fa970b63-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z\" (UID: \"d3f48375-4d0b-4876-b4e8-64a9fa970b63\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.822881 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3f48375-4d0b-4876-b4e8-64a9fa970b63-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z\" (UID: \"d3f48375-4d0b-4876-b4e8-64a9fa970b63\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.840789 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5wdk\" (UniqueName: \"kubernetes.io/projected/d3f48375-4d0b-4876-b4e8-64a9fa970b63-kube-api-access-g5wdk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z\" (UID: \"d3f48375-4d0b-4876-b4e8-64a9fa970b63\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" Nov 22 05:31:42 crc kubenswrapper[4818]: I1122 05:31:42.923426 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" Nov 22 05:31:43 crc kubenswrapper[4818]: I1122 05:31:43.506854 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z"] Nov 22 05:31:44 crc kubenswrapper[4818]: I1122 05:31:44.539134 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" event={"ID":"d3f48375-4d0b-4876-b4e8-64a9fa970b63","Type":"ContainerStarted","Data":"eebdb2c8bfddd2773bcc1202ddbc4f4d7cba44433a439b5e8621b8db4ebac3cd"} Nov 22 05:31:44 crc kubenswrapper[4818]: I1122 05:31:44.539621 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" event={"ID":"d3f48375-4d0b-4876-b4e8-64a9fa970b63","Type":"ContainerStarted","Data":"6c849ff9182922a5d3fa41d6d12af6e8581c9ae3f177dbd38c92c43f7747c384"} Nov 22 05:31:44 crc kubenswrapper[4818]: I1122 05:31:44.582380 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" podStartSLOduration=2.18201491 podStartE2EDuration="2.582356649s" podCreationTimestamp="2025-11-22 05:31:42 +0000 UTC" firstStartedPulling="2025-11-22 05:31:43.528710709 +0000 UTC m=+2656.103127276" lastFinishedPulling="2025-11-22 05:31:43.929052448 +0000 UTC m=+2656.503469015" observedRunningTime="2025-11-22 05:31:44.563767499 +0000 UTC m=+2657.138184066" watchObservedRunningTime="2025-11-22 05:31:44.582356649 +0000 UTC m=+2657.156773206" Nov 22 05:31:53 crc kubenswrapper[4818]: I1122 05:31:53.291177 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:31:53 crc kubenswrapper[4818]: I1122 05:31:53.653552 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"f9db7ffdf47d73a4c4ddb22a4b18596722d5f83c458d83cf3ee1c52ecf00e748"} Nov 22 05:32:35 crc kubenswrapper[4818]: I1122 05:32:35.114695 4818 generic.go:334] "Generic (PLEG): container finished" podID="d3f48375-4d0b-4876-b4e8-64a9fa970b63" containerID="eebdb2c8bfddd2773bcc1202ddbc4f4d7cba44433a439b5e8621b8db4ebac3cd" exitCode=0 Nov 22 05:32:35 crc kubenswrapper[4818]: I1122 05:32:35.114739 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" event={"ID":"d3f48375-4d0b-4876-b4e8-64a9fa970b63","Type":"ContainerDied","Data":"eebdb2c8bfddd2773bcc1202ddbc4f4d7cba44433a439b5e8621b8db4ebac3cd"} Nov 22 05:32:36 crc kubenswrapper[4818]: I1122 05:32:36.571070 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" Nov 22 05:32:36 crc kubenswrapper[4818]: I1122 05:32:36.587796 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5wdk\" (UniqueName: \"kubernetes.io/projected/d3f48375-4d0b-4876-b4e8-64a9fa970b63-kube-api-access-g5wdk\") pod \"d3f48375-4d0b-4876-b4e8-64a9fa970b63\" (UID: \"d3f48375-4d0b-4876-b4e8-64a9fa970b63\") " Nov 22 05:32:36 crc kubenswrapper[4818]: I1122 05:32:36.588410 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d3f48375-4d0b-4876-b4e8-64a9fa970b63-ceph\") pod \"d3f48375-4d0b-4876-b4e8-64a9fa970b63\" (UID: \"d3f48375-4d0b-4876-b4e8-64a9fa970b63\") " Nov 22 05:32:36 crc kubenswrapper[4818]: I1122 05:32:36.588509 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3f48375-4d0b-4876-b4e8-64a9fa970b63-ssh-key\") pod \"d3f48375-4d0b-4876-b4e8-64a9fa970b63\" (UID: \"d3f48375-4d0b-4876-b4e8-64a9fa970b63\") " Nov 22 05:32:36 crc kubenswrapper[4818]: I1122 05:32:36.588588 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3f48375-4d0b-4876-b4e8-64a9fa970b63-inventory\") pod \"d3f48375-4d0b-4876-b4e8-64a9fa970b63\" (UID: \"d3f48375-4d0b-4876-b4e8-64a9fa970b63\") " Nov 22 05:32:36 crc kubenswrapper[4818]: I1122 05:32:36.593866 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3f48375-4d0b-4876-b4e8-64a9fa970b63-kube-api-access-g5wdk" (OuterVolumeSpecName: "kube-api-access-g5wdk") pod "d3f48375-4d0b-4876-b4e8-64a9fa970b63" (UID: "d3f48375-4d0b-4876-b4e8-64a9fa970b63"). InnerVolumeSpecName "kube-api-access-g5wdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:32:36 crc kubenswrapper[4818]: I1122 05:32:36.598546 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3f48375-4d0b-4876-b4e8-64a9fa970b63-ceph" (OuterVolumeSpecName: "ceph") pod "d3f48375-4d0b-4876-b4e8-64a9fa970b63" (UID: "d3f48375-4d0b-4876-b4e8-64a9fa970b63"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:32:36 crc kubenswrapper[4818]: I1122 05:32:36.612242 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3f48375-4d0b-4876-b4e8-64a9fa970b63-inventory" (OuterVolumeSpecName: "inventory") pod "d3f48375-4d0b-4876-b4e8-64a9fa970b63" (UID: "d3f48375-4d0b-4876-b4e8-64a9fa970b63"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:32:36 crc kubenswrapper[4818]: I1122 05:32:36.629562 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3f48375-4d0b-4876-b4e8-64a9fa970b63-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d3f48375-4d0b-4876-b4e8-64a9fa970b63" (UID: "d3f48375-4d0b-4876-b4e8-64a9fa970b63"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:32:36 crc kubenswrapper[4818]: I1122 05:32:36.691129 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5wdk\" (UniqueName: \"kubernetes.io/projected/d3f48375-4d0b-4876-b4e8-64a9fa970b63-kube-api-access-g5wdk\") on node \"crc\" DevicePath \"\"" Nov 22 05:32:36 crc kubenswrapper[4818]: I1122 05:32:36.691183 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d3f48375-4d0b-4876-b4e8-64a9fa970b63-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:32:36 crc kubenswrapper[4818]: I1122 05:32:36.691202 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3f48375-4d0b-4876-b4e8-64a9fa970b63-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:32:36 crc kubenswrapper[4818]: I1122 05:32:36.691219 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3f48375-4d0b-4876-b4e8-64a9fa970b63-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.138814 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" event={"ID":"d3f48375-4d0b-4876-b4e8-64a9fa970b63","Type":"ContainerDied","Data":"6c849ff9182922a5d3fa41d6d12af6e8581c9ae3f177dbd38c92c43f7747c384"} Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.138863 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c849ff9182922a5d3fa41d6d12af6e8581c9ae3f177dbd38c92c43f7747c384" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.138892 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.267110 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-bg6f2"] Nov 22 05:32:37 crc kubenswrapper[4818]: E1122 05:32:37.268000 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f48375-4d0b-4876-b4e8-64a9fa970b63" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.268148 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f48375-4d0b-4876-b4e8-64a9fa970b63" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.268673 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3f48375-4d0b-4876-b4e8-64a9fa970b63" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.269786 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.273439 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.273924 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.275123 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.280048 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-bg6f2"] Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.281126 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.281887 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.409749 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/376ee003-816b-45fa-a57d-b9d2fa8120eb-ceph\") pod \"ssh-known-hosts-edpm-deployment-bg6f2\" (UID: \"376ee003-816b-45fa-a57d-b9d2fa8120eb\") " pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.411289 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/376ee003-816b-45fa-a57d-b9d2fa8120eb-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-bg6f2\" (UID: \"376ee003-816b-45fa-a57d-b9d2fa8120eb\") " pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.412385 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/376ee003-816b-45fa-a57d-b9d2fa8120eb-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-bg6f2\" (UID: \"376ee003-816b-45fa-a57d-b9d2fa8120eb\") " pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.412549 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hthsj\" (UniqueName: \"kubernetes.io/projected/376ee003-816b-45fa-a57d-b9d2fa8120eb-kube-api-access-hthsj\") pod \"ssh-known-hosts-edpm-deployment-bg6f2\" (UID: \"376ee003-816b-45fa-a57d-b9d2fa8120eb\") " pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.515078 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/376ee003-816b-45fa-a57d-b9d2fa8120eb-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-bg6f2\" (UID: \"376ee003-816b-45fa-a57d-b9d2fa8120eb\") " pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.515305 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/376ee003-816b-45fa-a57d-b9d2fa8120eb-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-bg6f2\" (UID: \"376ee003-816b-45fa-a57d-b9d2fa8120eb\") " pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.515387 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hthsj\" (UniqueName: \"kubernetes.io/projected/376ee003-816b-45fa-a57d-b9d2fa8120eb-kube-api-access-hthsj\") pod \"ssh-known-hosts-edpm-deployment-bg6f2\" (UID: \"376ee003-816b-45fa-a57d-b9d2fa8120eb\") " pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.515445 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/376ee003-816b-45fa-a57d-b9d2fa8120eb-ceph\") pod \"ssh-known-hosts-edpm-deployment-bg6f2\" (UID: \"376ee003-816b-45fa-a57d-b9d2fa8120eb\") " pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.523466 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/376ee003-816b-45fa-a57d-b9d2fa8120eb-ceph\") pod \"ssh-known-hosts-edpm-deployment-bg6f2\" (UID: \"376ee003-816b-45fa-a57d-b9d2fa8120eb\") " pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.524380 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/376ee003-816b-45fa-a57d-b9d2fa8120eb-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-bg6f2\" (UID: \"376ee003-816b-45fa-a57d-b9d2fa8120eb\") " pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.527526 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/376ee003-816b-45fa-a57d-b9d2fa8120eb-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-bg6f2\" (UID: \"376ee003-816b-45fa-a57d-b9d2fa8120eb\") " pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.549870 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hthsj\" (UniqueName: \"kubernetes.io/projected/376ee003-816b-45fa-a57d-b9d2fa8120eb-kube-api-access-hthsj\") pod \"ssh-known-hosts-edpm-deployment-bg6f2\" (UID: \"376ee003-816b-45fa-a57d-b9d2fa8120eb\") " pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" Nov 22 05:32:37 crc kubenswrapper[4818]: I1122 05:32:37.593015 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" Nov 22 05:32:38 crc kubenswrapper[4818]: I1122 05:32:38.256006 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-bg6f2"] Nov 22 05:32:38 crc kubenswrapper[4818]: W1122 05:32:38.269441 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod376ee003_816b_45fa_a57d_b9d2fa8120eb.slice/crio-2f58bf54eb419f509fb0ca3407b4509c28dc45d072bfb80c402f4d4d68bf0d3b WatchSource:0}: Error finding container 2f58bf54eb419f509fb0ca3407b4509c28dc45d072bfb80c402f4d4d68bf0d3b: Status 404 returned error can't find the container with id 2f58bf54eb419f509fb0ca3407b4509c28dc45d072bfb80c402f4d4d68bf0d3b Nov 22 05:32:39 crc kubenswrapper[4818]: I1122 05:32:39.159177 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" event={"ID":"376ee003-816b-45fa-a57d-b9d2fa8120eb","Type":"ContainerStarted","Data":"570dc46a76210589eb07c44a49afef4417146fc22d7dff7b16eb165533a8df13"} Nov 22 05:32:39 crc kubenswrapper[4818]: I1122 05:32:39.159572 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" event={"ID":"376ee003-816b-45fa-a57d-b9d2fa8120eb","Type":"ContainerStarted","Data":"2f58bf54eb419f509fb0ca3407b4509c28dc45d072bfb80c402f4d4d68bf0d3b"} Nov 22 05:32:39 crc kubenswrapper[4818]: I1122 05:32:39.176675 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" podStartSLOduration=1.665998699 podStartE2EDuration="2.176663648s" podCreationTimestamp="2025-11-22 05:32:37 +0000 UTC" firstStartedPulling="2025-11-22 05:32:38.271924938 +0000 UTC m=+2710.846341505" lastFinishedPulling="2025-11-22 05:32:38.782589887 +0000 UTC m=+2711.357006454" observedRunningTime="2025-11-22 05:32:39.176546175 +0000 UTC m=+2711.750962752" watchObservedRunningTime="2025-11-22 05:32:39.176663648 +0000 UTC m=+2711.751080175" Nov 22 05:32:50 crc kubenswrapper[4818]: I1122 05:32:50.267107 4818 generic.go:334] "Generic (PLEG): container finished" podID="376ee003-816b-45fa-a57d-b9d2fa8120eb" containerID="570dc46a76210589eb07c44a49afef4417146fc22d7dff7b16eb165533a8df13" exitCode=0 Nov 22 05:32:50 crc kubenswrapper[4818]: I1122 05:32:50.267280 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" event={"ID":"376ee003-816b-45fa-a57d-b9d2fa8120eb","Type":"ContainerDied","Data":"570dc46a76210589eb07c44a49afef4417146fc22d7dff7b16eb165533a8df13"} Nov 22 05:32:51 crc kubenswrapper[4818]: I1122 05:32:51.808350 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" Nov 22 05:32:51 crc kubenswrapper[4818]: I1122 05:32:51.925021 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hthsj\" (UniqueName: \"kubernetes.io/projected/376ee003-816b-45fa-a57d-b9d2fa8120eb-kube-api-access-hthsj\") pod \"376ee003-816b-45fa-a57d-b9d2fa8120eb\" (UID: \"376ee003-816b-45fa-a57d-b9d2fa8120eb\") " Nov 22 05:32:51 crc kubenswrapper[4818]: I1122 05:32:51.925170 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/376ee003-816b-45fa-a57d-b9d2fa8120eb-ssh-key-openstack-edpm-ipam\") pod \"376ee003-816b-45fa-a57d-b9d2fa8120eb\" (UID: \"376ee003-816b-45fa-a57d-b9d2fa8120eb\") " Nov 22 05:32:51 crc kubenswrapper[4818]: I1122 05:32:51.925228 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/376ee003-816b-45fa-a57d-b9d2fa8120eb-inventory-0\") pod \"376ee003-816b-45fa-a57d-b9d2fa8120eb\" (UID: \"376ee003-816b-45fa-a57d-b9d2fa8120eb\") " Nov 22 05:32:51 crc kubenswrapper[4818]: I1122 05:32:51.925331 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/376ee003-816b-45fa-a57d-b9d2fa8120eb-ceph\") pod \"376ee003-816b-45fa-a57d-b9d2fa8120eb\" (UID: \"376ee003-816b-45fa-a57d-b9d2fa8120eb\") " Nov 22 05:32:51 crc kubenswrapper[4818]: I1122 05:32:51.933131 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/376ee003-816b-45fa-a57d-b9d2fa8120eb-ceph" (OuterVolumeSpecName: "ceph") pod "376ee003-816b-45fa-a57d-b9d2fa8120eb" (UID: "376ee003-816b-45fa-a57d-b9d2fa8120eb"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:32:51 crc kubenswrapper[4818]: I1122 05:32:51.933521 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/376ee003-816b-45fa-a57d-b9d2fa8120eb-kube-api-access-hthsj" (OuterVolumeSpecName: "kube-api-access-hthsj") pod "376ee003-816b-45fa-a57d-b9d2fa8120eb" (UID: "376ee003-816b-45fa-a57d-b9d2fa8120eb"). InnerVolumeSpecName "kube-api-access-hthsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:32:51 crc kubenswrapper[4818]: I1122 05:32:51.977709 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/376ee003-816b-45fa-a57d-b9d2fa8120eb-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "376ee003-816b-45fa-a57d-b9d2fa8120eb" (UID: "376ee003-816b-45fa-a57d-b9d2fa8120eb"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:32:51 crc kubenswrapper[4818]: I1122 05:32:51.981051 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/376ee003-816b-45fa-a57d-b9d2fa8120eb-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "376ee003-816b-45fa-a57d-b9d2fa8120eb" (UID: "376ee003-816b-45fa-a57d-b9d2fa8120eb"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.027155 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hthsj\" (UniqueName: \"kubernetes.io/projected/376ee003-816b-45fa-a57d-b9d2fa8120eb-kube-api-access-hthsj\") on node \"crc\" DevicePath \"\"" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.027184 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/376ee003-816b-45fa-a57d-b9d2fa8120eb-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.027197 4818 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/376ee003-816b-45fa-a57d-b9d2fa8120eb-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.027209 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/376ee003-816b-45fa-a57d-b9d2fa8120eb-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.295940 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.309040 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-bg6f2" event={"ID":"376ee003-816b-45fa-a57d-b9d2fa8120eb","Type":"ContainerDied","Data":"2f58bf54eb419f509fb0ca3407b4509c28dc45d072bfb80c402f4d4d68bf0d3b"} Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.309087 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f58bf54eb419f509fb0ca3407b4509c28dc45d072bfb80c402f4d4d68bf0d3b" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.406052 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757"] Nov 22 05:32:52 crc kubenswrapper[4818]: E1122 05:32:52.406664 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="376ee003-816b-45fa-a57d-b9d2fa8120eb" containerName="ssh-known-hosts-edpm-deployment" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.406694 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="376ee003-816b-45fa-a57d-b9d2fa8120eb" containerName="ssh-known-hosts-edpm-deployment" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.407033 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="376ee003-816b-45fa-a57d-b9d2fa8120eb" containerName="ssh-known-hosts-edpm-deployment" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.409019 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.412892 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.413321 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.416190 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.416582 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.417039 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.441203 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757"] Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.537523 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8b757\" (UID: \"5d2cfc6e-9af5-4110-add7-b03c44ffbd33\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.537709 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h82m7\" (UniqueName: \"kubernetes.io/projected/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-kube-api-access-h82m7\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8b757\" (UID: \"5d2cfc6e-9af5-4110-add7-b03c44ffbd33\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.537757 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8b757\" (UID: \"5d2cfc6e-9af5-4110-add7-b03c44ffbd33\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.537965 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8b757\" (UID: \"5d2cfc6e-9af5-4110-add7-b03c44ffbd33\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.640044 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h82m7\" (UniqueName: \"kubernetes.io/projected/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-kube-api-access-h82m7\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8b757\" (UID: \"5d2cfc6e-9af5-4110-add7-b03c44ffbd33\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.640187 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8b757\" (UID: \"5d2cfc6e-9af5-4110-add7-b03c44ffbd33\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.640353 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8b757\" (UID: \"5d2cfc6e-9af5-4110-add7-b03c44ffbd33\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.641104 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8b757\" (UID: \"5d2cfc6e-9af5-4110-add7-b03c44ffbd33\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.645665 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8b757\" (UID: \"5d2cfc6e-9af5-4110-add7-b03c44ffbd33\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.646824 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8b757\" (UID: \"5d2cfc6e-9af5-4110-add7-b03c44ffbd33\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.647675 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8b757\" (UID: \"5d2cfc6e-9af5-4110-add7-b03c44ffbd33\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.672814 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h82m7\" (UniqueName: \"kubernetes.io/projected/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-kube-api-access-h82m7\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8b757\" (UID: \"5d2cfc6e-9af5-4110-add7-b03c44ffbd33\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" Nov 22 05:32:52 crc kubenswrapper[4818]: I1122 05:32:52.737377 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" Nov 22 05:32:53 crc kubenswrapper[4818]: I1122 05:32:53.402614 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757"] Nov 22 05:32:54 crc kubenswrapper[4818]: I1122 05:32:54.320057 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" event={"ID":"5d2cfc6e-9af5-4110-add7-b03c44ffbd33","Type":"ContainerStarted","Data":"7a6d52e802e7c29bf13802fa7175225ea0a0d762da34599219a5eabafb13814b"} Nov 22 05:32:54 crc kubenswrapper[4818]: I1122 05:32:54.320791 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" event={"ID":"5d2cfc6e-9af5-4110-add7-b03c44ffbd33","Type":"ContainerStarted","Data":"dda10fc1b445296db31748cf87429360bf0d9e3826c29347403c2dca3b81fdb0"} Nov 22 05:32:54 crc kubenswrapper[4818]: I1122 05:32:54.348091 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" podStartSLOduration=1.8713013630000002 podStartE2EDuration="2.348058969s" podCreationTimestamp="2025-11-22 05:32:52 +0000 UTC" firstStartedPulling="2025-11-22 05:32:53.399188102 +0000 UTC m=+2725.973604669" lastFinishedPulling="2025-11-22 05:32:53.875945708 +0000 UTC m=+2726.450362275" observedRunningTime="2025-11-22 05:32:54.344427782 +0000 UTC m=+2726.918844319" watchObservedRunningTime="2025-11-22 05:32:54.348058969 +0000 UTC m=+2726.922475556" Nov 22 05:33:02 crc kubenswrapper[4818]: I1122 05:33:02.403356 4818 generic.go:334] "Generic (PLEG): container finished" podID="5d2cfc6e-9af5-4110-add7-b03c44ffbd33" containerID="7a6d52e802e7c29bf13802fa7175225ea0a0d762da34599219a5eabafb13814b" exitCode=0 Nov 22 05:33:02 crc kubenswrapper[4818]: I1122 05:33:02.403425 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" event={"ID":"5d2cfc6e-9af5-4110-add7-b03c44ffbd33","Type":"ContainerDied","Data":"7a6d52e802e7c29bf13802fa7175225ea0a0d762da34599219a5eabafb13814b"} Nov 22 05:33:03 crc kubenswrapper[4818]: I1122 05:33:03.899732 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" Nov 22 05:33:03 crc kubenswrapper[4818]: I1122 05:33:03.971642 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h82m7\" (UniqueName: \"kubernetes.io/projected/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-kube-api-access-h82m7\") pod \"5d2cfc6e-9af5-4110-add7-b03c44ffbd33\" (UID: \"5d2cfc6e-9af5-4110-add7-b03c44ffbd33\") " Nov 22 05:33:03 crc kubenswrapper[4818]: I1122 05:33:03.971846 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-ssh-key\") pod \"5d2cfc6e-9af5-4110-add7-b03c44ffbd33\" (UID: \"5d2cfc6e-9af5-4110-add7-b03c44ffbd33\") " Nov 22 05:33:03 crc kubenswrapper[4818]: I1122 05:33:03.971930 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-ceph\") pod \"5d2cfc6e-9af5-4110-add7-b03c44ffbd33\" (UID: \"5d2cfc6e-9af5-4110-add7-b03c44ffbd33\") " Nov 22 05:33:03 crc kubenswrapper[4818]: I1122 05:33:03.971985 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-inventory\") pod \"5d2cfc6e-9af5-4110-add7-b03c44ffbd33\" (UID: \"5d2cfc6e-9af5-4110-add7-b03c44ffbd33\") " Nov 22 05:33:03 crc kubenswrapper[4818]: I1122 05:33:03.983094 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-ceph" (OuterVolumeSpecName: "ceph") pod "5d2cfc6e-9af5-4110-add7-b03c44ffbd33" (UID: "5d2cfc6e-9af5-4110-add7-b03c44ffbd33"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:33:03 crc kubenswrapper[4818]: I1122 05:33:03.983195 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-kube-api-access-h82m7" (OuterVolumeSpecName: "kube-api-access-h82m7") pod "5d2cfc6e-9af5-4110-add7-b03c44ffbd33" (UID: "5d2cfc6e-9af5-4110-add7-b03c44ffbd33"). InnerVolumeSpecName "kube-api-access-h82m7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.012003 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-inventory" (OuterVolumeSpecName: "inventory") pod "5d2cfc6e-9af5-4110-add7-b03c44ffbd33" (UID: "5d2cfc6e-9af5-4110-add7-b03c44ffbd33"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.023511 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5d2cfc6e-9af5-4110-add7-b03c44ffbd33" (UID: "5d2cfc6e-9af5-4110-add7-b03c44ffbd33"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.074925 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h82m7\" (UniqueName: \"kubernetes.io/projected/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-kube-api-access-h82m7\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.074972 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.074987 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.075022 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d2cfc6e-9af5-4110-add7-b03c44ffbd33-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.425166 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" event={"ID":"5d2cfc6e-9af5-4110-add7-b03c44ffbd33","Type":"ContainerDied","Data":"dda10fc1b445296db31748cf87429360bf0d9e3826c29347403c2dca3b81fdb0"} Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.425215 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dda10fc1b445296db31748cf87429360bf0d9e3826c29347403c2dca3b81fdb0" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.425337 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8b757" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.519456 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx"] Nov 22 05:33:04 crc kubenswrapper[4818]: E1122 05:33:04.519946 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d2cfc6e-9af5-4110-add7-b03c44ffbd33" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.519971 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d2cfc6e-9af5-4110-add7-b03c44ffbd33" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.520276 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d2cfc6e-9af5-4110-add7-b03c44ffbd33" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.520923 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.523758 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.523971 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.524231 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.524434 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.524969 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.528456 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx"] Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.686147 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx\" (UID: \"a072b598-ba0f-4b0d-8978-2d6e824ff5f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.686366 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s7f9\" (UniqueName: \"kubernetes.io/projected/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-kube-api-access-6s7f9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx\" (UID: \"a072b598-ba0f-4b0d-8978-2d6e824ff5f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.686449 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx\" (UID: \"a072b598-ba0f-4b0d-8978-2d6e824ff5f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.686527 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx\" (UID: \"a072b598-ba0f-4b0d-8978-2d6e824ff5f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.789353 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx\" (UID: \"a072b598-ba0f-4b0d-8978-2d6e824ff5f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.789508 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s7f9\" (UniqueName: \"kubernetes.io/projected/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-kube-api-access-6s7f9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx\" (UID: \"a072b598-ba0f-4b0d-8978-2d6e824ff5f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.789581 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx\" (UID: \"a072b598-ba0f-4b0d-8978-2d6e824ff5f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.789669 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx\" (UID: \"a072b598-ba0f-4b0d-8978-2d6e824ff5f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.794784 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx\" (UID: \"a072b598-ba0f-4b0d-8978-2d6e824ff5f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.795661 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx\" (UID: \"a072b598-ba0f-4b0d-8978-2d6e824ff5f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.796249 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx\" (UID: \"a072b598-ba0f-4b0d-8978-2d6e824ff5f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.815803 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s7f9\" (UniqueName: \"kubernetes.io/projected/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-kube-api-access-6s7f9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx\" (UID: \"a072b598-ba0f-4b0d-8978-2d6e824ff5f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" Nov 22 05:33:04 crc kubenswrapper[4818]: I1122 05:33:04.886129 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" Nov 22 05:33:05 crc kubenswrapper[4818]: I1122 05:33:05.451175 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx"] Nov 22 05:33:06 crc kubenswrapper[4818]: I1122 05:33:06.445270 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" event={"ID":"a072b598-ba0f-4b0d-8978-2d6e824ff5f6","Type":"ContainerStarted","Data":"e018c1bcef5f30dc85dd496ae911efb9bc64c16151313fbe1fdd5174736f39c5"} Nov 22 05:33:06 crc kubenswrapper[4818]: I1122 05:33:06.445620 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" event={"ID":"a072b598-ba0f-4b0d-8978-2d6e824ff5f6","Type":"ContainerStarted","Data":"fdcd962331bc8f07f383011ddeb2527e3b2a5fa064a3569dc3c34ce38660f648"} Nov 22 05:33:06 crc kubenswrapper[4818]: I1122 05:33:06.484833 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" podStartSLOduration=2.007556504 podStartE2EDuration="2.484814404s" podCreationTimestamp="2025-11-22 05:33:04 +0000 UTC" firstStartedPulling="2025-11-22 05:33:05.459886188 +0000 UTC m=+2738.034302725" lastFinishedPulling="2025-11-22 05:33:05.937144048 +0000 UTC m=+2738.511560625" observedRunningTime="2025-11-22 05:33:06.481847154 +0000 UTC m=+2739.056263681" watchObservedRunningTime="2025-11-22 05:33:06.484814404 +0000 UTC m=+2739.059230931" Nov 22 05:33:16 crc kubenswrapper[4818]: I1122 05:33:16.564452 4818 generic.go:334] "Generic (PLEG): container finished" podID="a072b598-ba0f-4b0d-8978-2d6e824ff5f6" containerID="e018c1bcef5f30dc85dd496ae911efb9bc64c16151313fbe1fdd5174736f39c5" exitCode=0 Nov 22 05:33:16 crc kubenswrapper[4818]: I1122 05:33:16.564661 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" event={"ID":"a072b598-ba0f-4b0d-8978-2d6e824ff5f6","Type":"ContainerDied","Data":"e018c1bcef5f30dc85dd496ae911efb9bc64c16151313fbe1fdd5174736f39c5"} Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.099206 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.216514 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6s7f9\" (UniqueName: \"kubernetes.io/projected/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-kube-api-access-6s7f9\") pod \"a072b598-ba0f-4b0d-8978-2d6e824ff5f6\" (UID: \"a072b598-ba0f-4b0d-8978-2d6e824ff5f6\") " Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.216598 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-ceph\") pod \"a072b598-ba0f-4b0d-8978-2d6e824ff5f6\" (UID: \"a072b598-ba0f-4b0d-8978-2d6e824ff5f6\") " Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.216673 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-inventory\") pod \"a072b598-ba0f-4b0d-8978-2d6e824ff5f6\" (UID: \"a072b598-ba0f-4b0d-8978-2d6e824ff5f6\") " Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.216707 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-ssh-key\") pod \"a072b598-ba0f-4b0d-8978-2d6e824ff5f6\" (UID: \"a072b598-ba0f-4b0d-8978-2d6e824ff5f6\") " Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.225580 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-ceph" (OuterVolumeSpecName: "ceph") pod "a072b598-ba0f-4b0d-8978-2d6e824ff5f6" (UID: "a072b598-ba0f-4b0d-8978-2d6e824ff5f6"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.228768 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-kube-api-access-6s7f9" (OuterVolumeSpecName: "kube-api-access-6s7f9") pod "a072b598-ba0f-4b0d-8978-2d6e824ff5f6" (UID: "a072b598-ba0f-4b0d-8978-2d6e824ff5f6"). InnerVolumeSpecName "kube-api-access-6s7f9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.293060 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a072b598-ba0f-4b0d-8978-2d6e824ff5f6" (UID: "a072b598-ba0f-4b0d-8978-2d6e824ff5f6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.311411 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-inventory" (OuterVolumeSpecName: "inventory") pod "a072b598-ba0f-4b0d-8978-2d6e824ff5f6" (UID: "a072b598-ba0f-4b0d-8978-2d6e824ff5f6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.319644 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.319701 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6s7f9\" (UniqueName: \"kubernetes.io/projected/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-kube-api-access-6s7f9\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.319716 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.319728 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a072b598-ba0f-4b0d-8978-2d6e824ff5f6-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.586880 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" event={"ID":"a072b598-ba0f-4b0d-8978-2d6e824ff5f6","Type":"ContainerDied","Data":"fdcd962331bc8f07f383011ddeb2527e3b2a5fa064a3569dc3c34ce38660f648"} Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.586931 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdcd962331bc8f07f383011ddeb2527e3b2a5fa064a3569dc3c34ce38660f648" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.587011 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.715627 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722"] Nov 22 05:33:18 crc kubenswrapper[4818]: E1122 05:33:18.715967 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a072b598-ba0f-4b0d-8978-2d6e824ff5f6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.715982 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a072b598-ba0f-4b0d-8978-2d6e824ff5f6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.716164 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a072b598-ba0f-4b0d-8978-2d6e824ff5f6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.716743 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.720775 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.720802 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.720935 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.721009 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.721104 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.721201 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.721523 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.723173 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.725935 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.725981 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.726011 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.726042 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.726085 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.726109 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnq29\" (UniqueName: \"kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-kube-api-access-qnq29\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.726146 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.726200 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.726221 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.726266 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.726316 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.726338 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.726391 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.737098 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722"] Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.827871 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.827956 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.828048 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.828085 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnq29\" (UniqueName: \"kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-kube-api-access-qnq29\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.828134 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.828216 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.828246 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.828314 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.828385 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.828427 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.828508 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.828948 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.828988 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.835470 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.835915 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.838129 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.838889 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.838891 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.838986 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.839606 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.840299 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.841055 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.843012 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.844485 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.844580 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:18 crc kubenswrapper[4818]: I1122 05:33:18.850768 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnq29\" (UniqueName: \"kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-kube-api-access-qnq29\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6h722\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:19 crc kubenswrapper[4818]: I1122 05:33:19.040406 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:19 crc kubenswrapper[4818]: I1122 05:33:19.434743 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722"] Nov 22 05:33:19 crc kubenswrapper[4818]: I1122 05:33:19.599866 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" event={"ID":"69bfd3ff-b747-43c0-82ab-ced23f3bc502","Type":"ContainerStarted","Data":"f562216adbef17ba9300425610fcafc3838fd588f88412f9ab27a4a5b0ce0839"} Nov 22 05:33:20 crc kubenswrapper[4818]: I1122 05:33:20.610209 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" event={"ID":"69bfd3ff-b747-43c0-82ab-ced23f3bc502","Type":"ContainerStarted","Data":"ba991a5135084859e1d68869307d66df2e0036e3aadbe15c18635c3eeb304eb9"} Nov 22 05:33:20 crc kubenswrapper[4818]: I1122 05:33:20.638537 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" podStartSLOduration=2.20763542 podStartE2EDuration="2.638509992s" podCreationTimestamp="2025-11-22 05:33:18 +0000 UTC" firstStartedPulling="2025-11-22 05:33:19.446375214 +0000 UTC m=+2752.020791741" lastFinishedPulling="2025-11-22 05:33:19.877249776 +0000 UTC m=+2752.451666313" observedRunningTime="2025-11-22 05:33:20.63173512 +0000 UTC m=+2753.206151647" watchObservedRunningTime="2025-11-22 05:33:20.638509992 +0000 UTC m=+2753.212926549" Nov 22 05:33:20 crc kubenswrapper[4818]: I1122 05:33:20.842129 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g5q2x"] Nov 22 05:33:20 crc kubenswrapper[4818]: I1122 05:33:20.846067 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5q2x" Nov 22 05:33:20 crc kubenswrapper[4818]: I1122 05:33:20.849109 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g5q2x"] Nov 22 05:33:20 crc kubenswrapper[4818]: I1122 05:33:20.970749 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ff5e15-ca46-4578-8fb7-32c2b0f33549-utilities\") pod \"redhat-operators-g5q2x\" (UID: \"b2ff5e15-ca46-4578-8fb7-32c2b0f33549\") " pod="openshift-marketplace/redhat-operators-g5q2x" Nov 22 05:33:20 crc kubenswrapper[4818]: I1122 05:33:20.970897 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ff5e15-ca46-4578-8fb7-32c2b0f33549-catalog-content\") pod \"redhat-operators-g5q2x\" (UID: \"b2ff5e15-ca46-4578-8fb7-32c2b0f33549\") " pod="openshift-marketplace/redhat-operators-g5q2x" Nov 22 05:33:20 crc kubenswrapper[4818]: I1122 05:33:20.970934 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnlns\" (UniqueName: \"kubernetes.io/projected/b2ff5e15-ca46-4578-8fb7-32c2b0f33549-kube-api-access-vnlns\") pod \"redhat-operators-g5q2x\" (UID: \"b2ff5e15-ca46-4578-8fb7-32c2b0f33549\") " pod="openshift-marketplace/redhat-operators-g5q2x" Nov 22 05:33:21 crc kubenswrapper[4818]: I1122 05:33:21.072657 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ff5e15-ca46-4578-8fb7-32c2b0f33549-utilities\") pod \"redhat-operators-g5q2x\" (UID: \"b2ff5e15-ca46-4578-8fb7-32c2b0f33549\") " pod="openshift-marketplace/redhat-operators-g5q2x" Nov 22 05:33:21 crc kubenswrapper[4818]: I1122 05:33:21.072907 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ff5e15-ca46-4578-8fb7-32c2b0f33549-catalog-content\") pod \"redhat-operators-g5q2x\" (UID: \"b2ff5e15-ca46-4578-8fb7-32c2b0f33549\") " pod="openshift-marketplace/redhat-operators-g5q2x" Nov 22 05:33:21 crc kubenswrapper[4818]: I1122 05:33:21.072964 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnlns\" (UniqueName: \"kubernetes.io/projected/b2ff5e15-ca46-4578-8fb7-32c2b0f33549-kube-api-access-vnlns\") pod \"redhat-operators-g5q2x\" (UID: \"b2ff5e15-ca46-4578-8fb7-32c2b0f33549\") " pod="openshift-marketplace/redhat-operators-g5q2x" Nov 22 05:33:21 crc kubenswrapper[4818]: I1122 05:33:21.073245 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ff5e15-ca46-4578-8fb7-32c2b0f33549-catalog-content\") pod \"redhat-operators-g5q2x\" (UID: \"b2ff5e15-ca46-4578-8fb7-32c2b0f33549\") " pod="openshift-marketplace/redhat-operators-g5q2x" Nov 22 05:33:21 crc kubenswrapper[4818]: I1122 05:33:21.073517 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ff5e15-ca46-4578-8fb7-32c2b0f33549-utilities\") pod \"redhat-operators-g5q2x\" (UID: \"b2ff5e15-ca46-4578-8fb7-32c2b0f33549\") " pod="openshift-marketplace/redhat-operators-g5q2x" Nov 22 05:33:21 crc kubenswrapper[4818]: I1122 05:33:21.094267 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnlns\" (UniqueName: \"kubernetes.io/projected/b2ff5e15-ca46-4578-8fb7-32c2b0f33549-kube-api-access-vnlns\") pod \"redhat-operators-g5q2x\" (UID: \"b2ff5e15-ca46-4578-8fb7-32c2b0f33549\") " pod="openshift-marketplace/redhat-operators-g5q2x" Nov 22 05:33:21 crc kubenswrapper[4818]: I1122 05:33:21.168888 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5q2x" Nov 22 05:33:21 crc kubenswrapper[4818]: I1122 05:33:21.624271 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g5q2x"] Nov 22 05:33:22 crc kubenswrapper[4818]: I1122 05:33:22.631905 4818 generic.go:334] "Generic (PLEG): container finished" podID="b2ff5e15-ca46-4578-8fb7-32c2b0f33549" containerID="3548191b355a4cd6b806cfd8e6896e1bbd94dcf45e6c23b1fbd1193222d2b183" exitCode=0 Nov 22 05:33:22 crc kubenswrapper[4818]: I1122 05:33:22.632243 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5q2x" event={"ID":"b2ff5e15-ca46-4578-8fb7-32c2b0f33549","Type":"ContainerDied","Data":"3548191b355a4cd6b806cfd8e6896e1bbd94dcf45e6c23b1fbd1193222d2b183"} Nov 22 05:33:22 crc kubenswrapper[4818]: I1122 05:33:22.632318 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5q2x" event={"ID":"b2ff5e15-ca46-4578-8fb7-32c2b0f33549","Type":"ContainerStarted","Data":"46f79498dae229e5b10c96fadd21932ecaa99e5d28886f8cc309479ccc49591d"} Nov 22 05:33:23 crc kubenswrapper[4818]: I1122 05:33:23.645331 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5q2x" event={"ID":"b2ff5e15-ca46-4578-8fb7-32c2b0f33549","Type":"ContainerStarted","Data":"c347ed09a56b642a2801072b3c9ae699549d88d9d30f8bbef2ed673415502093"} Nov 22 05:33:24 crc kubenswrapper[4818]: I1122 05:33:24.663475 4818 generic.go:334] "Generic (PLEG): container finished" podID="b2ff5e15-ca46-4578-8fb7-32c2b0f33549" containerID="c347ed09a56b642a2801072b3c9ae699549d88d9d30f8bbef2ed673415502093" exitCode=0 Nov 22 05:33:24 crc kubenswrapper[4818]: I1122 05:33:24.663539 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5q2x" event={"ID":"b2ff5e15-ca46-4578-8fb7-32c2b0f33549","Type":"ContainerDied","Data":"c347ed09a56b642a2801072b3c9ae699549d88d9d30f8bbef2ed673415502093"} Nov 22 05:33:25 crc kubenswrapper[4818]: I1122 05:33:25.677445 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5q2x" event={"ID":"b2ff5e15-ca46-4578-8fb7-32c2b0f33549","Type":"ContainerStarted","Data":"51174dfffe3c563f5406add68a9289efa3f1b2acb43d70715fe0dee0cd007a33"} Nov 22 05:33:25 crc kubenswrapper[4818]: I1122 05:33:25.713540 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g5q2x" podStartSLOduration=3.246893223 podStartE2EDuration="5.713516527s" podCreationTimestamp="2025-11-22 05:33:20 +0000 UTC" firstStartedPulling="2025-11-22 05:33:22.635595314 +0000 UTC m=+2755.210011881" lastFinishedPulling="2025-11-22 05:33:25.102218618 +0000 UTC m=+2757.676635185" observedRunningTime="2025-11-22 05:33:25.703888858 +0000 UTC m=+2758.278305425" watchObservedRunningTime="2025-11-22 05:33:25.713516527 +0000 UTC m=+2758.287933094" Nov 22 05:33:31 crc kubenswrapper[4818]: I1122 05:33:31.169205 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g5q2x" Nov 22 05:33:31 crc kubenswrapper[4818]: I1122 05:33:31.169802 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g5q2x" Nov 22 05:33:32 crc kubenswrapper[4818]: I1122 05:33:32.248447 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-g5q2x" podUID="b2ff5e15-ca46-4578-8fb7-32c2b0f33549" containerName="registry-server" probeResult="failure" output=< Nov 22 05:33:32 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Nov 22 05:33:32 crc kubenswrapper[4818]: > Nov 22 05:33:41 crc kubenswrapper[4818]: I1122 05:33:41.220605 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g5q2x" Nov 22 05:33:41 crc kubenswrapper[4818]: I1122 05:33:41.276855 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g5q2x" Nov 22 05:33:41 crc kubenswrapper[4818]: I1122 05:33:41.459864 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g5q2x"] Nov 22 05:33:42 crc kubenswrapper[4818]: I1122 05:33:42.849515 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g5q2x" podUID="b2ff5e15-ca46-4578-8fb7-32c2b0f33549" containerName="registry-server" containerID="cri-o://51174dfffe3c563f5406add68a9289efa3f1b2acb43d70715fe0dee0cd007a33" gracePeriod=2 Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.370378 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5q2x" Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.463593 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnlns\" (UniqueName: \"kubernetes.io/projected/b2ff5e15-ca46-4578-8fb7-32c2b0f33549-kube-api-access-vnlns\") pod \"b2ff5e15-ca46-4578-8fb7-32c2b0f33549\" (UID: \"b2ff5e15-ca46-4578-8fb7-32c2b0f33549\") " Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.463997 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ff5e15-ca46-4578-8fb7-32c2b0f33549-catalog-content\") pod \"b2ff5e15-ca46-4578-8fb7-32c2b0f33549\" (UID: \"b2ff5e15-ca46-4578-8fb7-32c2b0f33549\") " Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.464120 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ff5e15-ca46-4578-8fb7-32c2b0f33549-utilities\") pod \"b2ff5e15-ca46-4578-8fb7-32c2b0f33549\" (UID: \"b2ff5e15-ca46-4578-8fb7-32c2b0f33549\") " Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.464892 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2ff5e15-ca46-4578-8fb7-32c2b0f33549-utilities" (OuterVolumeSpecName: "utilities") pod "b2ff5e15-ca46-4578-8fb7-32c2b0f33549" (UID: "b2ff5e15-ca46-4578-8fb7-32c2b0f33549"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.468992 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2ff5e15-ca46-4578-8fb7-32c2b0f33549-kube-api-access-vnlns" (OuterVolumeSpecName: "kube-api-access-vnlns") pod "b2ff5e15-ca46-4578-8fb7-32c2b0f33549" (UID: "b2ff5e15-ca46-4578-8fb7-32c2b0f33549"). InnerVolumeSpecName "kube-api-access-vnlns". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.545057 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2ff5e15-ca46-4578-8fb7-32c2b0f33549-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2ff5e15-ca46-4578-8fb7-32c2b0f33549" (UID: "b2ff5e15-ca46-4578-8fb7-32c2b0f33549"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.566371 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ff5e15-ca46-4578-8fb7-32c2b0f33549-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.566415 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ff5e15-ca46-4578-8fb7-32c2b0f33549-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.566427 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnlns\" (UniqueName: \"kubernetes.io/projected/b2ff5e15-ca46-4578-8fb7-32c2b0f33549-kube-api-access-vnlns\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.865308 4818 generic.go:334] "Generic (PLEG): container finished" podID="b2ff5e15-ca46-4578-8fb7-32c2b0f33549" containerID="51174dfffe3c563f5406add68a9289efa3f1b2acb43d70715fe0dee0cd007a33" exitCode=0 Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.865375 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5q2x" event={"ID":"b2ff5e15-ca46-4578-8fb7-32c2b0f33549","Type":"ContainerDied","Data":"51174dfffe3c563f5406add68a9289efa3f1b2acb43d70715fe0dee0cd007a33"} Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.865418 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5q2x" Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.865425 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5q2x" event={"ID":"b2ff5e15-ca46-4578-8fb7-32c2b0f33549","Type":"ContainerDied","Data":"46f79498dae229e5b10c96fadd21932ecaa99e5d28886f8cc309479ccc49591d"} Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.865473 4818 scope.go:117] "RemoveContainer" containerID="51174dfffe3c563f5406add68a9289efa3f1b2acb43d70715fe0dee0cd007a33" Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.901981 4818 scope.go:117] "RemoveContainer" containerID="c347ed09a56b642a2801072b3c9ae699549d88d9d30f8bbef2ed673415502093" Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.928214 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g5q2x"] Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.941999 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g5q2x"] Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.949149 4818 scope.go:117] "RemoveContainer" containerID="3548191b355a4cd6b806cfd8e6896e1bbd94dcf45e6c23b1fbd1193222d2b183" Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.985561 4818 scope.go:117] "RemoveContainer" containerID="51174dfffe3c563f5406add68a9289efa3f1b2acb43d70715fe0dee0cd007a33" Nov 22 05:33:43 crc kubenswrapper[4818]: E1122 05:33:43.986189 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51174dfffe3c563f5406add68a9289efa3f1b2acb43d70715fe0dee0cd007a33\": container with ID starting with 51174dfffe3c563f5406add68a9289efa3f1b2acb43d70715fe0dee0cd007a33 not found: ID does not exist" containerID="51174dfffe3c563f5406add68a9289efa3f1b2acb43d70715fe0dee0cd007a33" Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.986246 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51174dfffe3c563f5406add68a9289efa3f1b2acb43d70715fe0dee0cd007a33"} err="failed to get container status \"51174dfffe3c563f5406add68a9289efa3f1b2acb43d70715fe0dee0cd007a33\": rpc error: code = NotFound desc = could not find container \"51174dfffe3c563f5406add68a9289efa3f1b2acb43d70715fe0dee0cd007a33\": container with ID starting with 51174dfffe3c563f5406add68a9289efa3f1b2acb43d70715fe0dee0cd007a33 not found: ID does not exist" Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.986310 4818 scope.go:117] "RemoveContainer" containerID="c347ed09a56b642a2801072b3c9ae699549d88d9d30f8bbef2ed673415502093" Nov 22 05:33:43 crc kubenswrapper[4818]: E1122 05:33:43.986677 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c347ed09a56b642a2801072b3c9ae699549d88d9d30f8bbef2ed673415502093\": container with ID starting with c347ed09a56b642a2801072b3c9ae699549d88d9d30f8bbef2ed673415502093 not found: ID does not exist" containerID="c347ed09a56b642a2801072b3c9ae699549d88d9d30f8bbef2ed673415502093" Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.986722 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c347ed09a56b642a2801072b3c9ae699549d88d9d30f8bbef2ed673415502093"} err="failed to get container status \"c347ed09a56b642a2801072b3c9ae699549d88d9d30f8bbef2ed673415502093\": rpc error: code = NotFound desc = could not find container \"c347ed09a56b642a2801072b3c9ae699549d88d9d30f8bbef2ed673415502093\": container with ID starting with c347ed09a56b642a2801072b3c9ae699549d88d9d30f8bbef2ed673415502093 not found: ID does not exist" Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.986749 4818 scope.go:117] "RemoveContainer" containerID="3548191b355a4cd6b806cfd8e6896e1bbd94dcf45e6c23b1fbd1193222d2b183" Nov 22 05:33:43 crc kubenswrapper[4818]: E1122 05:33:43.987244 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3548191b355a4cd6b806cfd8e6896e1bbd94dcf45e6c23b1fbd1193222d2b183\": container with ID starting with 3548191b355a4cd6b806cfd8e6896e1bbd94dcf45e6c23b1fbd1193222d2b183 not found: ID does not exist" containerID="3548191b355a4cd6b806cfd8e6896e1bbd94dcf45e6c23b1fbd1193222d2b183" Nov 22 05:33:43 crc kubenswrapper[4818]: I1122 05:33:43.987292 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3548191b355a4cd6b806cfd8e6896e1bbd94dcf45e6c23b1fbd1193222d2b183"} err="failed to get container status \"3548191b355a4cd6b806cfd8e6896e1bbd94dcf45e6c23b1fbd1193222d2b183\": rpc error: code = NotFound desc = could not find container \"3548191b355a4cd6b806cfd8e6896e1bbd94dcf45e6c23b1fbd1193222d2b183\": container with ID starting with 3548191b355a4cd6b806cfd8e6896e1bbd94dcf45e6c23b1fbd1193222d2b183 not found: ID does not exist" Nov 22 05:33:44 crc kubenswrapper[4818]: I1122 05:33:44.303870 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2ff5e15-ca46-4578-8fb7-32c2b0f33549" path="/var/lib/kubelet/pods/b2ff5e15-ca46-4578-8fb7-32c2b0f33549/volumes" Nov 22 05:33:58 crc kubenswrapper[4818]: I1122 05:33:58.026678 4818 generic.go:334] "Generic (PLEG): container finished" podID="69bfd3ff-b747-43c0-82ab-ced23f3bc502" containerID="ba991a5135084859e1d68869307d66df2e0036e3aadbe15c18635c3eeb304eb9" exitCode=0 Nov 22 05:33:58 crc kubenswrapper[4818]: I1122 05:33:58.026790 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" event={"ID":"69bfd3ff-b747-43c0-82ab-ced23f3bc502","Type":"ContainerDied","Data":"ba991a5135084859e1d68869307d66df2e0036e3aadbe15c18635c3eeb304eb9"} Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.530109 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.629038 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-neutron-metadata-combined-ca-bundle\") pod \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.630199 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.630390 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-bootstrap-combined-ca-bundle\") pod \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.630470 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-ssh-key\") pod \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.630535 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.630581 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-ovn-combined-ca-bundle\") pod \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.630627 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnq29\" (UniqueName: \"kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-kube-api-access-qnq29\") pod \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.630808 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-openstack-edpm-ipam-ovn-default-certs-0\") pod \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.630941 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-nova-combined-ca-bundle\") pod \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.631010 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-ceph\") pod \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.631122 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-repo-setup-combined-ca-bundle\") pod \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.631246 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-libvirt-combined-ca-bundle\") pod \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.631560 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-inventory\") pod \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\" (UID: \"69bfd3ff-b747-43c0-82ab-ced23f3bc502\") " Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.637188 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "69bfd3ff-b747-43c0-82ab-ced23f3bc502" (UID: "69bfd3ff-b747-43c0-82ab-ced23f3bc502"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.637357 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "69bfd3ff-b747-43c0-82ab-ced23f3bc502" (UID: "69bfd3ff-b747-43c0-82ab-ced23f3bc502"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.637744 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "69bfd3ff-b747-43c0-82ab-ced23f3bc502" (UID: "69bfd3ff-b747-43c0-82ab-ced23f3bc502"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.638627 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "69bfd3ff-b747-43c0-82ab-ced23f3bc502" (UID: "69bfd3ff-b747-43c0-82ab-ced23f3bc502"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.639106 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "69bfd3ff-b747-43c0-82ab-ced23f3bc502" (UID: "69bfd3ff-b747-43c0-82ab-ced23f3bc502"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.640453 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "69bfd3ff-b747-43c0-82ab-ced23f3bc502" (UID: "69bfd3ff-b747-43c0-82ab-ced23f3bc502"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.640755 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-kube-api-access-qnq29" (OuterVolumeSpecName: "kube-api-access-qnq29") pod "69bfd3ff-b747-43c0-82ab-ced23f3bc502" (UID: "69bfd3ff-b747-43c0-82ab-ced23f3bc502"). InnerVolumeSpecName "kube-api-access-qnq29". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.642562 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "69bfd3ff-b747-43c0-82ab-ced23f3bc502" (UID: "69bfd3ff-b747-43c0-82ab-ced23f3bc502"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.643506 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "69bfd3ff-b747-43c0-82ab-ced23f3bc502" (UID: "69bfd3ff-b747-43c0-82ab-ced23f3bc502"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.651462 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "69bfd3ff-b747-43c0-82ab-ced23f3bc502" (UID: "69bfd3ff-b747-43c0-82ab-ced23f3bc502"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.651517 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-ceph" (OuterVolumeSpecName: "ceph") pod "69bfd3ff-b747-43c0-82ab-ced23f3bc502" (UID: "69bfd3ff-b747-43c0-82ab-ced23f3bc502"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.677160 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-inventory" (OuterVolumeSpecName: "inventory") pod "69bfd3ff-b747-43c0-82ab-ced23f3bc502" (UID: "69bfd3ff-b747-43c0-82ab-ced23f3bc502"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.680525 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "69bfd3ff-b747-43c0-82ab-ced23f3bc502" (UID: "69bfd3ff-b747-43c0-82ab-ced23f3bc502"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.735195 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.735717 4818 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.735841 4818 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.735960 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnq29\" (UniqueName: \"kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-kube-api-access-qnq29\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.736080 4818 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.736297 4818 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.736467 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.736662 4818 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.736848 4818 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.736970 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.737078 4818 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.737187 4818 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/69bfd3ff-b747-43c0-82ab-ced23f3bc502-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 22 05:33:59 crc kubenswrapper[4818]: I1122 05:33:59.737350 4818 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69bfd3ff-b747-43c0-82ab-ced23f3bc502-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.050195 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.050122 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6h722" event={"ID":"69bfd3ff-b747-43c0-82ab-ced23f3bc502","Type":"ContainerDied","Data":"f562216adbef17ba9300425610fcafc3838fd588f88412f9ab27a4a5b0ce0839"} Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.054499 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f562216adbef17ba9300425610fcafc3838fd588f88412f9ab27a4a5b0ce0839" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.191818 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2"] Nov 22 05:34:00 crc kubenswrapper[4818]: E1122 05:34:00.192224 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ff5e15-ca46-4578-8fb7-32c2b0f33549" containerName="extract-content" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.192240 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ff5e15-ca46-4578-8fb7-32c2b0f33549" containerName="extract-content" Nov 22 05:34:00 crc kubenswrapper[4818]: E1122 05:34:00.192279 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ff5e15-ca46-4578-8fb7-32c2b0f33549" containerName="registry-server" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.192287 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ff5e15-ca46-4578-8fb7-32c2b0f33549" containerName="registry-server" Nov 22 05:34:00 crc kubenswrapper[4818]: E1122 05:34:00.192300 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ff5e15-ca46-4578-8fb7-32c2b0f33549" containerName="extract-utilities" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.192307 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ff5e15-ca46-4578-8fb7-32c2b0f33549" containerName="extract-utilities" Nov 22 05:34:00 crc kubenswrapper[4818]: E1122 05:34:00.192322 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69bfd3ff-b747-43c0-82ab-ced23f3bc502" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.192329 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="69bfd3ff-b747-43c0-82ab-ced23f3bc502" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.192515 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2ff5e15-ca46-4578-8fb7-32c2b0f33549" containerName="registry-server" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.192529 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="69bfd3ff-b747-43c0-82ab-ced23f3bc502" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.193154 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.195374 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.195577 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.195578 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.195870 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.195989 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.199677 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2"] Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.245751 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1488ee6c-884b-4a61-93c9-ae582c9f4778-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2\" (UID: \"1488ee6c-884b-4a61-93c9-ae582c9f4778\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.245859 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1488ee6c-884b-4a61-93c9-ae582c9f4778-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2\" (UID: \"1488ee6c-884b-4a61-93c9-ae582c9f4778\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.246176 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1488ee6c-884b-4a61-93c9-ae582c9f4778-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2\" (UID: \"1488ee6c-884b-4a61-93c9-ae582c9f4778\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.246244 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fr98m\" (UniqueName: \"kubernetes.io/projected/1488ee6c-884b-4a61-93c9-ae582c9f4778-kube-api-access-fr98m\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2\" (UID: \"1488ee6c-884b-4a61-93c9-ae582c9f4778\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.347843 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1488ee6c-884b-4a61-93c9-ae582c9f4778-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2\" (UID: \"1488ee6c-884b-4a61-93c9-ae582c9f4778\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.348051 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1488ee6c-884b-4a61-93c9-ae582c9f4778-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2\" (UID: \"1488ee6c-884b-4a61-93c9-ae582c9f4778\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.348105 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fr98m\" (UniqueName: \"kubernetes.io/projected/1488ee6c-884b-4a61-93c9-ae582c9f4778-kube-api-access-fr98m\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2\" (UID: \"1488ee6c-884b-4a61-93c9-ae582c9f4778\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.348176 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1488ee6c-884b-4a61-93c9-ae582c9f4778-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2\" (UID: \"1488ee6c-884b-4a61-93c9-ae582c9f4778\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.354541 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1488ee6c-884b-4a61-93c9-ae582c9f4778-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2\" (UID: \"1488ee6c-884b-4a61-93c9-ae582c9f4778\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.354745 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1488ee6c-884b-4a61-93c9-ae582c9f4778-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2\" (UID: \"1488ee6c-884b-4a61-93c9-ae582c9f4778\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.356874 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1488ee6c-884b-4a61-93c9-ae582c9f4778-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2\" (UID: \"1488ee6c-884b-4a61-93c9-ae582c9f4778\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.364937 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fr98m\" (UniqueName: \"kubernetes.io/projected/1488ee6c-884b-4a61-93c9-ae582c9f4778-kube-api-access-fr98m\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2\" (UID: \"1488ee6c-884b-4a61-93c9-ae582c9f4778\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" Nov 22 05:34:00 crc kubenswrapper[4818]: I1122 05:34:00.507563 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" Nov 22 05:34:01 crc kubenswrapper[4818]: I1122 05:34:01.059204 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2"] Nov 22 05:34:02 crc kubenswrapper[4818]: I1122 05:34:02.074764 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" event={"ID":"1488ee6c-884b-4a61-93c9-ae582c9f4778","Type":"ContainerStarted","Data":"895583ab9571029d7311ca834110dabdbfd1be14a6fd97c49fbffe82511b2522"} Nov 22 05:34:02 crc kubenswrapper[4818]: I1122 05:34:02.075031 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" event={"ID":"1488ee6c-884b-4a61-93c9-ae582c9f4778","Type":"ContainerStarted","Data":"ed4510df175f027f95827dbf16b509a46209fc5b9962faf78771137e91445d3a"} Nov 22 05:34:02 crc kubenswrapper[4818]: I1122 05:34:02.096553 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" podStartSLOduration=1.639056426 podStartE2EDuration="2.096528913s" podCreationTimestamp="2025-11-22 05:34:00 +0000 UTC" firstStartedPulling="2025-11-22 05:34:01.072029919 +0000 UTC m=+2793.646446456" lastFinishedPulling="2025-11-22 05:34:01.529502416 +0000 UTC m=+2794.103918943" observedRunningTime="2025-11-22 05:34:02.089747551 +0000 UTC m=+2794.664164078" watchObservedRunningTime="2025-11-22 05:34:02.096528913 +0000 UTC m=+2794.670945440" Nov 22 05:34:02 crc kubenswrapper[4818]: I1122 05:34:02.242019 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k82v7"] Nov 22 05:34:02 crc kubenswrapper[4818]: I1122 05:34:02.244569 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k82v7" Nov 22 05:34:02 crc kubenswrapper[4818]: I1122 05:34:02.250838 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k82v7"] Nov 22 05:34:02 crc kubenswrapper[4818]: I1122 05:34:02.280607 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e3b7701-7441-4328-8599-d5086a4f850b-catalog-content\") pod \"certified-operators-k82v7\" (UID: \"6e3b7701-7441-4328-8599-d5086a4f850b\") " pod="openshift-marketplace/certified-operators-k82v7" Nov 22 05:34:02 crc kubenswrapper[4818]: I1122 05:34:02.280789 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcpcv\" (UniqueName: \"kubernetes.io/projected/6e3b7701-7441-4328-8599-d5086a4f850b-kube-api-access-vcpcv\") pod \"certified-operators-k82v7\" (UID: \"6e3b7701-7441-4328-8599-d5086a4f850b\") " pod="openshift-marketplace/certified-operators-k82v7" Nov 22 05:34:02 crc kubenswrapper[4818]: I1122 05:34:02.280827 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e3b7701-7441-4328-8599-d5086a4f850b-utilities\") pod \"certified-operators-k82v7\" (UID: \"6e3b7701-7441-4328-8599-d5086a4f850b\") " pod="openshift-marketplace/certified-operators-k82v7" Nov 22 05:34:02 crc kubenswrapper[4818]: I1122 05:34:02.382417 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcpcv\" (UniqueName: \"kubernetes.io/projected/6e3b7701-7441-4328-8599-d5086a4f850b-kube-api-access-vcpcv\") pod \"certified-operators-k82v7\" (UID: \"6e3b7701-7441-4328-8599-d5086a4f850b\") " pod="openshift-marketplace/certified-operators-k82v7" Nov 22 05:34:02 crc kubenswrapper[4818]: I1122 05:34:02.382660 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e3b7701-7441-4328-8599-d5086a4f850b-utilities\") pod \"certified-operators-k82v7\" (UID: \"6e3b7701-7441-4328-8599-d5086a4f850b\") " pod="openshift-marketplace/certified-operators-k82v7" Nov 22 05:34:02 crc kubenswrapper[4818]: I1122 05:34:02.382795 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e3b7701-7441-4328-8599-d5086a4f850b-catalog-content\") pod \"certified-operators-k82v7\" (UID: \"6e3b7701-7441-4328-8599-d5086a4f850b\") " pod="openshift-marketplace/certified-operators-k82v7" Nov 22 05:34:02 crc kubenswrapper[4818]: I1122 05:34:02.383205 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e3b7701-7441-4328-8599-d5086a4f850b-catalog-content\") pod \"certified-operators-k82v7\" (UID: \"6e3b7701-7441-4328-8599-d5086a4f850b\") " pod="openshift-marketplace/certified-operators-k82v7" Nov 22 05:34:02 crc kubenswrapper[4818]: I1122 05:34:02.383302 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e3b7701-7441-4328-8599-d5086a4f850b-utilities\") pod \"certified-operators-k82v7\" (UID: \"6e3b7701-7441-4328-8599-d5086a4f850b\") " pod="openshift-marketplace/certified-operators-k82v7" Nov 22 05:34:02 crc kubenswrapper[4818]: I1122 05:34:02.401891 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcpcv\" (UniqueName: \"kubernetes.io/projected/6e3b7701-7441-4328-8599-d5086a4f850b-kube-api-access-vcpcv\") pod \"certified-operators-k82v7\" (UID: \"6e3b7701-7441-4328-8599-d5086a4f850b\") " pod="openshift-marketplace/certified-operators-k82v7" Nov 22 05:34:02 crc kubenswrapper[4818]: I1122 05:34:02.561055 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k82v7" Nov 22 05:34:03 crc kubenswrapper[4818]: I1122 05:34:03.108038 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k82v7"] Nov 22 05:34:03 crc kubenswrapper[4818]: W1122 05:34:03.110355 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e3b7701_7441_4328_8599_d5086a4f850b.slice/crio-f7756d14771d44365036dbfbcf8cc963ab6c271c0901083162f420ebc0e16677 WatchSource:0}: Error finding container f7756d14771d44365036dbfbcf8cc963ab6c271c0901083162f420ebc0e16677: Status 404 returned error can't find the container with id f7756d14771d44365036dbfbcf8cc963ab6c271c0901083162f420ebc0e16677 Nov 22 05:34:04 crc kubenswrapper[4818]: I1122 05:34:04.089705 4818 generic.go:334] "Generic (PLEG): container finished" podID="6e3b7701-7441-4328-8599-d5086a4f850b" containerID="d1628d5798b4b544f542a9d6374b109f67243b75b2dfc48b1b439d514ebed02c" exitCode=0 Nov 22 05:34:04 crc kubenswrapper[4818]: I1122 05:34:04.089763 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k82v7" event={"ID":"6e3b7701-7441-4328-8599-d5086a4f850b","Type":"ContainerDied","Data":"d1628d5798b4b544f542a9d6374b109f67243b75b2dfc48b1b439d514ebed02c"} Nov 22 05:34:04 crc kubenswrapper[4818]: I1122 05:34:04.090110 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k82v7" event={"ID":"6e3b7701-7441-4328-8599-d5086a4f850b","Type":"ContainerStarted","Data":"f7756d14771d44365036dbfbcf8cc963ab6c271c0901083162f420ebc0e16677"} Nov 22 05:34:05 crc kubenswrapper[4818]: I1122 05:34:05.105050 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k82v7" event={"ID":"6e3b7701-7441-4328-8599-d5086a4f850b","Type":"ContainerStarted","Data":"2bbaa29a5e7e74667f022f49aa4b1fd81dbf8af5d5bf0b2377aa8919ef72082b"} Nov 22 05:34:06 crc kubenswrapper[4818]: I1122 05:34:06.122352 4818 generic.go:334] "Generic (PLEG): container finished" podID="6e3b7701-7441-4328-8599-d5086a4f850b" containerID="2bbaa29a5e7e74667f022f49aa4b1fd81dbf8af5d5bf0b2377aa8919ef72082b" exitCode=0 Nov 22 05:34:06 crc kubenswrapper[4818]: I1122 05:34:06.122398 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k82v7" event={"ID":"6e3b7701-7441-4328-8599-d5086a4f850b","Type":"ContainerDied","Data":"2bbaa29a5e7e74667f022f49aa4b1fd81dbf8af5d5bf0b2377aa8919ef72082b"} Nov 22 05:34:07 crc kubenswrapper[4818]: I1122 05:34:07.133498 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k82v7" event={"ID":"6e3b7701-7441-4328-8599-d5086a4f850b","Type":"ContainerStarted","Data":"d24f754ccdf8457abba145186244bb1e7d01fefac89a2aa39130ef37b82d7d37"} Nov 22 05:34:07 crc kubenswrapper[4818]: I1122 05:34:07.159509 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k82v7" podStartSLOduration=2.727663716 podStartE2EDuration="5.159487594s" podCreationTimestamp="2025-11-22 05:34:02 +0000 UTC" firstStartedPulling="2025-11-22 05:34:04.092986488 +0000 UTC m=+2796.667403015" lastFinishedPulling="2025-11-22 05:34:06.524810366 +0000 UTC m=+2799.099226893" observedRunningTime="2025-11-22 05:34:07.152427774 +0000 UTC m=+2799.726844341" watchObservedRunningTime="2025-11-22 05:34:07.159487594 +0000 UTC m=+2799.733904131" Nov 22 05:34:08 crc kubenswrapper[4818]: I1122 05:34:08.141979 4818 generic.go:334] "Generic (PLEG): container finished" podID="1488ee6c-884b-4a61-93c9-ae582c9f4778" containerID="895583ab9571029d7311ca834110dabdbfd1be14a6fd97c49fbffe82511b2522" exitCode=0 Nov 22 05:34:08 crc kubenswrapper[4818]: I1122 05:34:08.142063 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" event={"ID":"1488ee6c-884b-4a61-93c9-ae582c9f4778","Type":"ContainerDied","Data":"895583ab9571029d7311ca834110dabdbfd1be14a6fd97c49fbffe82511b2522"} Nov 22 05:34:09 crc kubenswrapper[4818]: I1122 05:34:09.589158 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" Nov 22 05:34:09 crc kubenswrapper[4818]: I1122 05:34:09.738840 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1488ee6c-884b-4a61-93c9-ae582c9f4778-ceph\") pod \"1488ee6c-884b-4a61-93c9-ae582c9f4778\" (UID: \"1488ee6c-884b-4a61-93c9-ae582c9f4778\") " Nov 22 05:34:09 crc kubenswrapper[4818]: I1122 05:34:09.738941 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fr98m\" (UniqueName: \"kubernetes.io/projected/1488ee6c-884b-4a61-93c9-ae582c9f4778-kube-api-access-fr98m\") pod \"1488ee6c-884b-4a61-93c9-ae582c9f4778\" (UID: \"1488ee6c-884b-4a61-93c9-ae582c9f4778\") " Nov 22 05:34:09 crc kubenswrapper[4818]: I1122 05:34:09.739001 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1488ee6c-884b-4a61-93c9-ae582c9f4778-inventory\") pod \"1488ee6c-884b-4a61-93c9-ae582c9f4778\" (UID: \"1488ee6c-884b-4a61-93c9-ae582c9f4778\") " Nov 22 05:34:09 crc kubenswrapper[4818]: I1122 05:34:09.739163 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1488ee6c-884b-4a61-93c9-ae582c9f4778-ssh-key\") pod \"1488ee6c-884b-4a61-93c9-ae582c9f4778\" (UID: \"1488ee6c-884b-4a61-93c9-ae582c9f4778\") " Nov 22 05:34:09 crc kubenswrapper[4818]: I1122 05:34:09.745469 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1488ee6c-884b-4a61-93c9-ae582c9f4778-ceph" (OuterVolumeSpecName: "ceph") pod "1488ee6c-884b-4a61-93c9-ae582c9f4778" (UID: "1488ee6c-884b-4a61-93c9-ae582c9f4778"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:34:09 crc kubenswrapper[4818]: I1122 05:34:09.746640 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1488ee6c-884b-4a61-93c9-ae582c9f4778-kube-api-access-fr98m" (OuterVolumeSpecName: "kube-api-access-fr98m") pod "1488ee6c-884b-4a61-93c9-ae582c9f4778" (UID: "1488ee6c-884b-4a61-93c9-ae582c9f4778"). InnerVolumeSpecName "kube-api-access-fr98m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:34:09 crc kubenswrapper[4818]: I1122 05:34:09.775214 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1488ee6c-884b-4a61-93c9-ae582c9f4778-inventory" (OuterVolumeSpecName: "inventory") pod "1488ee6c-884b-4a61-93c9-ae582c9f4778" (UID: "1488ee6c-884b-4a61-93c9-ae582c9f4778"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:34:09 crc kubenswrapper[4818]: I1122 05:34:09.782807 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1488ee6c-884b-4a61-93c9-ae582c9f4778-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1488ee6c-884b-4a61-93c9-ae582c9f4778" (UID: "1488ee6c-884b-4a61-93c9-ae582c9f4778"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:34:09 crc kubenswrapper[4818]: I1122 05:34:09.840772 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1488ee6c-884b-4a61-93c9-ae582c9f4778-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:34:09 crc kubenswrapper[4818]: I1122 05:34:09.840806 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1488ee6c-884b-4a61-93c9-ae582c9f4778-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:34:09 crc kubenswrapper[4818]: I1122 05:34:09.840817 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fr98m\" (UniqueName: \"kubernetes.io/projected/1488ee6c-884b-4a61-93c9-ae582c9f4778-kube-api-access-fr98m\") on node \"crc\" DevicePath \"\"" Nov 22 05:34:09 crc kubenswrapper[4818]: I1122 05:34:09.840832 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1488ee6c-884b-4a61-93c9-ae582c9f4778-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.159137 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" event={"ID":"1488ee6c-884b-4a61-93c9-ae582c9f4778","Type":"ContainerDied","Data":"ed4510df175f027f95827dbf16b509a46209fc5b9962faf78771137e91445d3a"} Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.159207 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed4510df175f027f95827dbf16b509a46209fc5b9962faf78771137e91445d3a" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.159208 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.243700 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm"] Nov 22 05:34:10 crc kubenswrapper[4818]: E1122 05:34:10.244456 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1488ee6c-884b-4a61-93c9-ae582c9f4778" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.244504 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="1488ee6c-884b-4a61-93c9-ae582c9f4778" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.244956 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="1488ee6c-884b-4a61-93c9-ae582c9f4778" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.246100 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.248160 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.248588 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.248722 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.249085 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.249508 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.251194 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.252447 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm"] Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.349231 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k2rlm\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.349635 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k2rlm\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.349665 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k2rlm\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.349739 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znthj\" (UniqueName: \"kubernetes.io/projected/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-kube-api-access-znthj\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k2rlm\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.349803 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k2rlm\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.349846 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k2rlm\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.452202 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k2rlm\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.452333 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k2rlm\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.452437 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k2rlm\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.452471 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k2rlm\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.452579 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znthj\" (UniqueName: \"kubernetes.io/projected/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-kube-api-access-znthj\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k2rlm\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.453907 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k2rlm\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.454247 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k2rlm\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.458216 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k2rlm\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.459395 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k2rlm\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.459767 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k2rlm\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.461679 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k2rlm\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.474413 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znthj\" (UniqueName: \"kubernetes.io/projected/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-kube-api-access-znthj\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k2rlm\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.569852 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:34:10 crc kubenswrapper[4818]: I1122 05:34:10.980351 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm"] Nov 22 05:34:10 crc kubenswrapper[4818]: W1122 05:34:10.989703 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c86c5bf_d3d5_48b2_adfe_8da2c05d2cef.slice/crio-d93db8c9deaba62e859bb688db5086926d99180e181bb858471b0d4c420e2f7b WatchSource:0}: Error finding container d93db8c9deaba62e859bb688db5086926d99180e181bb858471b0d4c420e2f7b: Status 404 returned error can't find the container with id d93db8c9deaba62e859bb688db5086926d99180e181bb858471b0d4c420e2f7b Nov 22 05:34:11 crc kubenswrapper[4818]: I1122 05:34:11.176437 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" event={"ID":"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef","Type":"ContainerStarted","Data":"d93db8c9deaba62e859bb688db5086926d99180e181bb858471b0d4c420e2f7b"} Nov 22 05:34:12 crc kubenswrapper[4818]: I1122 05:34:12.194490 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" event={"ID":"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef","Type":"ContainerStarted","Data":"0093c38d2e638f5ecb6b4c33bbf355cff805ce0aff861da9bf1041dfd69ce172"} Nov 22 05:34:12 crc kubenswrapper[4818]: I1122 05:34:12.222611 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" podStartSLOduration=1.739853211 podStartE2EDuration="2.222589589s" podCreationTimestamp="2025-11-22 05:34:10 +0000 UTC" firstStartedPulling="2025-11-22 05:34:10.993039593 +0000 UTC m=+2803.567456140" lastFinishedPulling="2025-11-22 05:34:11.475775991 +0000 UTC m=+2804.050192518" observedRunningTime="2025-11-22 05:34:12.215983382 +0000 UTC m=+2804.790399929" watchObservedRunningTime="2025-11-22 05:34:12.222589589 +0000 UTC m=+2804.797006136" Nov 22 05:34:12 crc kubenswrapper[4818]: I1122 05:34:12.561916 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k82v7" Nov 22 05:34:12 crc kubenswrapper[4818]: I1122 05:34:12.562195 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k82v7" Nov 22 05:34:12 crc kubenswrapper[4818]: I1122 05:34:12.651442 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k82v7" Nov 22 05:34:13 crc kubenswrapper[4818]: I1122 05:34:13.272955 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k82v7" Nov 22 05:34:13 crc kubenswrapper[4818]: I1122 05:34:13.328424 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k82v7"] Nov 22 05:34:15 crc kubenswrapper[4818]: I1122 05:34:15.234057 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k82v7" podUID="6e3b7701-7441-4328-8599-d5086a4f850b" containerName="registry-server" containerID="cri-o://d24f754ccdf8457abba145186244bb1e7d01fefac89a2aa39130ef37b82d7d37" gracePeriod=2 Nov 22 05:34:15 crc kubenswrapper[4818]: I1122 05:34:15.743162 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k82v7" Nov 22 05:34:15 crc kubenswrapper[4818]: I1122 05:34:15.765592 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e3b7701-7441-4328-8599-d5086a4f850b-utilities\") pod \"6e3b7701-7441-4328-8599-d5086a4f850b\" (UID: \"6e3b7701-7441-4328-8599-d5086a4f850b\") " Nov 22 05:34:15 crc kubenswrapper[4818]: I1122 05:34:15.765923 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e3b7701-7441-4328-8599-d5086a4f850b-catalog-content\") pod \"6e3b7701-7441-4328-8599-d5086a4f850b\" (UID: \"6e3b7701-7441-4328-8599-d5086a4f850b\") " Nov 22 05:34:15 crc kubenswrapper[4818]: I1122 05:34:15.766075 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcpcv\" (UniqueName: \"kubernetes.io/projected/6e3b7701-7441-4328-8599-d5086a4f850b-kube-api-access-vcpcv\") pod \"6e3b7701-7441-4328-8599-d5086a4f850b\" (UID: \"6e3b7701-7441-4328-8599-d5086a4f850b\") " Nov 22 05:34:15 crc kubenswrapper[4818]: I1122 05:34:15.766500 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e3b7701-7441-4328-8599-d5086a4f850b-utilities" (OuterVolumeSpecName: "utilities") pod "6e3b7701-7441-4328-8599-d5086a4f850b" (UID: "6e3b7701-7441-4328-8599-d5086a4f850b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:34:15 crc kubenswrapper[4818]: I1122 05:34:15.766770 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e3b7701-7441-4328-8599-d5086a4f850b-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 05:34:15 crc kubenswrapper[4818]: I1122 05:34:15.774449 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e3b7701-7441-4328-8599-d5086a4f850b-kube-api-access-vcpcv" (OuterVolumeSpecName: "kube-api-access-vcpcv") pod "6e3b7701-7441-4328-8599-d5086a4f850b" (UID: "6e3b7701-7441-4328-8599-d5086a4f850b"). InnerVolumeSpecName "kube-api-access-vcpcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:34:15 crc kubenswrapper[4818]: I1122 05:34:15.818161 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e3b7701-7441-4328-8599-d5086a4f850b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e3b7701-7441-4328-8599-d5086a4f850b" (UID: "6e3b7701-7441-4328-8599-d5086a4f850b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:34:15 crc kubenswrapper[4818]: I1122 05:34:15.868137 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e3b7701-7441-4328-8599-d5086a4f850b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 05:34:15 crc kubenswrapper[4818]: I1122 05:34:15.868172 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcpcv\" (UniqueName: \"kubernetes.io/projected/6e3b7701-7441-4328-8599-d5086a4f850b-kube-api-access-vcpcv\") on node \"crc\" DevicePath \"\"" Nov 22 05:34:16 crc kubenswrapper[4818]: I1122 05:34:16.244736 4818 generic.go:334] "Generic (PLEG): container finished" podID="6e3b7701-7441-4328-8599-d5086a4f850b" containerID="d24f754ccdf8457abba145186244bb1e7d01fefac89a2aa39130ef37b82d7d37" exitCode=0 Nov 22 05:34:16 crc kubenswrapper[4818]: I1122 05:34:16.244785 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k82v7" event={"ID":"6e3b7701-7441-4328-8599-d5086a4f850b","Type":"ContainerDied","Data":"d24f754ccdf8457abba145186244bb1e7d01fefac89a2aa39130ef37b82d7d37"} Nov 22 05:34:16 crc kubenswrapper[4818]: I1122 05:34:16.244811 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k82v7" event={"ID":"6e3b7701-7441-4328-8599-d5086a4f850b","Type":"ContainerDied","Data":"f7756d14771d44365036dbfbcf8cc963ab6c271c0901083162f420ebc0e16677"} Nov 22 05:34:16 crc kubenswrapper[4818]: I1122 05:34:16.244832 4818 scope.go:117] "RemoveContainer" containerID="d24f754ccdf8457abba145186244bb1e7d01fefac89a2aa39130ef37b82d7d37" Nov 22 05:34:16 crc kubenswrapper[4818]: I1122 05:34:16.244831 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k82v7" Nov 22 05:34:16 crc kubenswrapper[4818]: I1122 05:34:16.290432 4818 scope.go:117] "RemoveContainer" containerID="2bbaa29a5e7e74667f022f49aa4b1fd81dbf8af5d5bf0b2377aa8919ef72082b" Nov 22 05:34:16 crc kubenswrapper[4818]: I1122 05:34:16.322460 4818 scope.go:117] "RemoveContainer" containerID="d1628d5798b4b544f542a9d6374b109f67243b75b2dfc48b1b439d514ebed02c" Nov 22 05:34:16 crc kubenswrapper[4818]: I1122 05:34:16.336537 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k82v7"] Nov 22 05:34:16 crc kubenswrapper[4818]: I1122 05:34:16.347150 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k82v7"] Nov 22 05:34:16 crc kubenswrapper[4818]: I1122 05:34:16.354539 4818 scope.go:117] "RemoveContainer" containerID="d24f754ccdf8457abba145186244bb1e7d01fefac89a2aa39130ef37b82d7d37" Nov 22 05:34:16 crc kubenswrapper[4818]: E1122 05:34:16.354929 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d24f754ccdf8457abba145186244bb1e7d01fefac89a2aa39130ef37b82d7d37\": container with ID starting with d24f754ccdf8457abba145186244bb1e7d01fefac89a2aa39130ef37b82d7d37 not found: ID does not exist" containerID="d24f754ccdf8457abba145186244bb1e7d01fefac89a2aa39130ef37b82d7d37" Nov 22 05:34:16 crc kubenswrapper[4818]: I1122 05:34:16.354957 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d24f754ccdf8457abba145186244bb1e7d01fefac89a2aa39130ef37b82d7d37"} err="failed to get container status \"d24f754ccdf8457abba145186244bb1e7d01fefac89a2aa39130ef37b82d7d37\": rpc error: code = NotFound desc = could not find container \"d24f754ccdf8457abba145186244bb1e7d01fefac89a2aa39130ef37b82d7d37\": container with ID starting with d24f754ccdf8457abba145186244bb1e7d01fefac89a2aa39130ef37b82d7d37 not found: ID does not exist" Nov 22 05:34:16 crc kubenswrapper[4818]: I1122 05:34:16.354979 4818 scope.go:117] "RemoveContainer" containerID="2bbaa29a5e7e74667f022f49aa4b1fd81dbf8af5d5bf0b2377aa8919ef72082b" Nov 22 05:34:16 crc kubenswrapper[4818]: E1122 05:34:16.355417 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bbaa29a5e7e74667f022f49aa4b1fd81dbf8af5d5bf0b2377aa8919ef72082b\": container with ID starting with 2bbaa29a5e7e74667f022f49aa4b1fd81dbf8af5d5bf0b2377aa8919ef72082b not found: ID does not exist" containerID="2bbaa29a5e7e74667f022f49aa4b1fd81dbf8af5d5bf0b2377aa8919ef72082b" Nov 22 05:34:16 crc kubenswrapper[4818]: I1122 05:34:16.355438 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bbaa29a5e7e74667f022f49aa4b1fd81dbf8af5d5bf0b2377aa8919ef72082b"} err="failed to get container status \"2bbaa29a5e7e74667f022f49aa4b1fd81dbf8af5d5bf0b2377aa8919ef72082b\": rpc error: code = NotFound desc = could not find container \"2bbaa29a5e7e74667f022f49aa4b1fd81dbf8af5d5bf0b2377aa8919ef72082b\": container with ID starting with 2bbaa29a5e7e74667f022f49aa4b1fd81dbf8af5d5bf0b2377aa8919ef72082b not found: ID does not exist" Nov 22 05:34:16 crc kubenswrapper[4818]: I1122 05:34:16.355451 4818 scope.go:117] "RemoveContainer" containerID="d1628d5798b4b544f542a9d6374b109f67243b75b2dfc48b1b439d514ebed02c" Nov 22 05:34:16 crc kubenswrapper[4818]: E1122 05:34:16.356362 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1628d5798b4b544f542a9d6374b109f67243b75b2dfc48b1b439d514ebed02c\": container with ID starting with d1628d5798b4b544f542a9d6374b109f67243b75b2dfc48b1b439d514ebed02c not found: ID does not exist" containerID="d1628d5798b4b544f542a9d6374b109f67243b75b2dfc48b1b439d514ebed02c" Nov 22 05:34:16 crc kubenswrapper[4818]: I1122 05:34:16.356457 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1628d5798b4b544f542a9d6374b109f67243b75b2dfc48b1b439d514ebed02c"} err="failed to get container status \"d1628d5798b4b544f542a9d6374b109f67243b75b2dfc48b1b439d514ebed02c\": rpc error: code = NotFound desc = could not find container \"d1628d5798b4b544f542a9d6374b109f67243b75b2dfc48b1b439d514ebed02c\": container with ID starting with d1628d5798b4b544f542a9d6374b109f67243b75b2dfc48b1b439d514ebed02c not found: ID does not exist" Nov 22 05:34:18 crc kubenswrapper[4818]: I1122 05:34:18.309063 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e3b7701-7441-4328-8599-d5086a4f850b" path="/var/lib/kubelet/pods/6e3b7701-7441-4328-8599-d5086a4f850b/volumes" Nov 22 05:34:21 crc kubenswrapper[4818]: I1122 05:34:21.265522 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:34:21 crc kubenswrapper[4818]: I1122 05:34:21.265940 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:34:36 crc kubenswrapper[4818]: I1122 05:34:36.368977 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mzs9b"] Nov 22 05:34:36 crc kubenswrapper[4818]: E1122 05:34:36.372546 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e3b7701-7441-4328-8599-d5086a4f850b" containerName="registry-server" Nov 22 05:34:36 crc kubenswrapper[4818]: I1122 05:34:36.372592 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e3b7701-7441-4328-8599-d5086a4f850b" containerName="registry-server" Nov 22 05:34:36 crc kubenswrapper[4818]: E1122 05:34:36.372607 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e3b7701-7441-4328-8599-d5086a4f850b" containerName="extract-content" Nov 22 05:34:36 crc kubenswrapper[4818]: I1122 05:34:36.372617 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e3b7701-7441-4328-8599-d5086a4f850b" containerName="extract-content" Nov 22 05:34:36 crc kubenswrapper[4818]: E1122 05:34:36.372630 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e3b7701-7441-4328-8599-d5086a4f850b" containerName="extract-utilities" Nov 22 05:34:36 crc kubenswrapper[4818]: I1122 05:34:36.372638 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e3b7701-7441-4328-8599-d5086a4f850b" containerName="extract-utilities" Nov 22 05:34:36 crc kubenswrapper[4818]: I1122 05:34:36.373351 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e3b7701-7441-4328-8599-d5086a4f850b" containerName="registry-server" Nov 22 05:34:36 crc kubenswrapper[4818]: I1122 05:34:36.375347 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mzs9b" Nov 22 05:34:36 crc kubenswrapper[4818]: I1122 05:34:36.401416 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mzs9b"] Nov 22 05:34:36 crc kubenswrapper[4818]: I1122 05:34:36.405265 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3109e10-d592-4ff7-b6d4-82e6f07307f4-catalog-content\") pod \"redhat-marketplace-mzs9b\" (UID: \"a3109e10-d592-4ff7-b6d4-82e6f07307f4\") " pod="openshift-marketplace/redhat-marketplace-mzs9b" Nov 22 05:34:36 crc kubenswrapper[4818]: I1122 05:34:36.405373 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgs25\" (UniqueName: \"kubernetes.io/projected/a3109e10-d592-4ff7-b6d4-82e6f07307f4-kube-api-access-kgs25\") pod \"redhat-marketplace-mzs9b\" (UID: \"a3109e10-d592-4ff7-b6d4-82e6f07307f4\") " pod="openshift-marketplace/redhat-marketplace-mzs9b" Nov 22 05:34:36 crc kubenswrapper[4818]: I1122 05:34:36.405404 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3109e10-d592-4ff7-b6d4-82e6f07307f4-utilities\") pod \"redhat-marketplace-mzs9b\" (UID: \"a3109e10-d592-4ff7-b6d4-82e6f07307f4\") " pod="openshift-marketplace/redhat-marketplace-mzs9b" Nov 22 05:34:36 crc kubenswrapper[4818]: I1122 05:34:36.507242 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3109e10-d592-4ff7-b6d4-82e6f07307f4-catalog-content\") pod \"redhat-marketplace-mzs9b\" (UID: \"a3109e10-d592-4ff7-b6d4-82e6f07307f4\") " pod="openshift-marketplace/redhat-marketplace-mzs9b" Nov 22 05:34:36 crc kubenswrapper[4818]: I1122 05:34:36.507331 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgs25\" (UniqueName: \"kubernetes.io/projected/a3109e10-d592-4ff7-b6d4-82e6f07307f4-kube-api-access-kgs25\") pod \"redhat-marketplace-mzs9b\" (UID: \"a3109e10-d592-4ff7-b6d4-82e6f07307f4\") " pod="openshift-marketplace/redhat-marketplace-mzs9b" Nov 22 05:34:36 crc kubenswrapper[4818]: I1122 05:34:36.507354 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3109e10-d592-4ff7-b6d4-82e6f07307f4-utilities\") pod \"redhat-marketplace-mzs9b\" (UID: \"a3109e10-d592-4ff7-b6d4-82e6f07307f4\") " pod="openshift-marketplace/redhat-marketplace-mzs9b" Nov 22 05:34:36 crc kubenswrapper[4818]: I1122 05:34:36.507974 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3109e10-d592-4ff7-b6d4-82e6f07307f4-utilities\") pod \"redhat-marketplace-mzs9b\" (UID: \"a3109e10-d592-4ff7-b6d4-82e6f07307f4\") " pod="openshift-marketplace/redhat-marketplace-mzs9b" Nov 22 05:34:36 crc kubenswrapper[4818]: I1122 05:34:36.508110 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3109e10-d592-4ff7-b6d4-82e6f07307f4-catalog-content\") pod \"redhat-marketplace-mzs9b\" (UID: \"a3109e10-d592-4ff7-b6d4-82e6f07307f4\") " pod="openshift-marketplace/redhat-marketplace-mzs9b" Nov 22 05:34:36 crc kubenswrapper[4818]: I1122 05:34:36.528684 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgs25\" (UniqueName: \"kubernetes.io/projected/a3109e10-d592-4ff7-b6d4-82e6f07307f4-kube-api-access-kgs25\") pod \"redhat-marketplace-mzs9b\" (UID: \"a3109e10-d592-4ff7-b6d4-82e6f07307f4\") " pod="openshift-marketplace/redhat-marketplace-mzs9b" Nov 22 05:34:36 crc kubenswrapper[4818]: I1122 05:34:36.711340 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mzs9b" Nov 22 05:34:37 crc kubenswrapper[4818]: I1122 05:34:37.182598 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mzs9b"] Nov 22 05:34:37 crc kubenswrapper[4818]: I1122 05:34:37.467068 4818 generic.go:334] "Generic (PLEG): container finished" podID="a3109e10-d592-4ff7-b6d4-82e6f07307f4" containerID="168f184c196aa909e9a197b9da56cd9270501525d8bcb9a05819b34d908abfe1" exitCode=0 Nov 22 05:34:37 crc kubenswrapper[4818]: I1122 05:34:37.467117 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzs9b" event={"ID":"a3109e10-d592-4ff7-b6d4-82e6f07307f4","Type":"ContainerDied","Data":"168f184c196aa909e9a197b9da56cd9270501525d8bcb9a05819b34d908abfe1"} Nov 22 05:34:37 crc kubenswrapper[4818]: I1122 05:34:37.467148 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzs9b" event={"ID":"a3109e10-d592-4ff7-b6d4-82e6f07307f4","Type":"ContainerStarted","Data":"91fa872c1b5540c1f7e938c7658625c22dba370c4bb562fa14a8c4caf7b88008"} Nov 22 05:34:38 crc kubenswrapper[4818]: I1122 05:34:38.476419 4818 generic.go:334] "Generic (PLEG): container finished" podID="a3109e10-d592-4ff7-b6d4-82e6f07307f4" containerID="4d57ee0b43f124ad876eb2672ca22620e5123c8cccdea3712b9a9d001a901782" exitCode=0 Nov 22 05:34:38 crc kubenswrapper[4818]: I1122 05:34:38.476535 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzs9b" event={"ID":"a3109e10-d592-4ff7-b6d4-82e6f07307f4","Type":"ContainerDied","Data":"4d57ee0b43f124ad876eb2672ca22620e5123c8cccdea3712b9a9d001a901782"} Nov 22 05:34:40 crc kubenswrapper[4818]: I1122 05:34:40.496596 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzs9b" event={"ID":"a3109e10-d592-4ff7-b6d4-82e6f07307f4","Type":"ContainerStarted","Data":"b77119f3bc17040655407e94dd17c5f8e7f2acb5601246d22d29121e6cc008ee"} Nov 22 05:34:40 crc kubenswrapper[4818]: I1122 05:34:40.519781 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mzs9b" podStartSLOduration=2.646000752 podStartE2EDuration="4.519762474s" podCreationTimestamp="2025-11-22 05:34:36 +0000 UTC" firstStartedPulling="2025-11-22 05:34:37.469851274 +0000 UTC m=+2830.044267811" lastFinishedPulling="2025-11-22 05:34:39.343612996 +0000 UTC m=+2831.918029533" observedRunningTime="2025-11-22 05:34:40.513971257 +0000 UTC m=+2833.088387794" watchObservedRunningTime="2025-11-22 05:34:40.519762474 +0000 UTC m=+2833.094179011" Nov 22 05:34:46 crc kubenswrapper[4818]: I1122 05:34:46.711988 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mzs9b" Nov 22 05:34:46 crc kubenswrapper[4818]: I1122 05:34:46.712641 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mzs9b" Nov 22 05:34:46 crc kubenswrapper[4818]: I1122 05:34:46.794958 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mzs9b" Nov 22 05:34:47 crc kubenswrapper[4818]: I1122 05:34:47.626217 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mzs9b" Nov 22 05:34:47 crc kubenswrapper[4818]: I1122 05:34:47.680058 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mzs9b"] Nov 22 05:34:49 crc kubenswrapper[4818]: I1122 05:34:49.578736 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mzs9b" podUID="a3109e10-d592-4ff7-b6d4-82e6f07307f4" containerName="registry-server" containerID="cri-o://b77119f3bc17040655407e94dd17c5f8e7f2acb5601246d22d29121e6cc008ee" gracePeriod=2 Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.066490 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mzs9b" Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.198444 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3109e10-d592-4ff7-b6d4-82e6f07307f4-catalog-content\") pod \"a3109e10-d592-4ff7-b6d4-82e6f07307f4\" (UID: \"a3109e10-d592-4ff7-b6d4-82e6f07307f4\") " Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.198584 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgs25\" (UniqueName: \"kubernetes.io/projected/a3109e10-d592-4ff7-b6d4-82e6f07307f4-kube-api-access-kgs25\") pod \"a3109e10-d592-4ff7-b6d4-82e6f07307f4\" (UID: \"a3109e10-d592-4ff7-b6d4-82e6f07307f4\") " Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.198697 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3109e10-d592-4ff7-b6d4-82e6f07307f4-utilities\") pod \"a3109e10-d592-4ff7-b6d4-82e6f07307f4\" (UID: \"a3109e10-d592-4ff7-b6d4-82e6f07307f4\") " Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.199622 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3109e10-d592-4ff7-b6d4-82e6f07307f4-utilities" (OuterVolumeSpecName: "utilities") pod "a3109e10-d592-4ff7-b6d4-82e6f07307f4" (UID: "a3109e10-d592-4ff7-b6d4-82e6f07307f4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.204071 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3109e10-d592-4ff7-b6d4-82e6f07307f4-kube-api-access-kgs25" (OuterVolumeSpecName: "kube-api-access-kgs25") pod "a3109e10-d592-4ff7-b6d4-82e6f07307f4" (UID: "a3109e10-d592-4ff7-b6d4-82e6f07307f4"). InnerVolumeSpecName "kube-api-access-kgs25". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.215045 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3109e10-d592-4ff7-b6d4-82e6f07307f4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a3109e10-d592-4ff7-b6d4-82e6f07307f4" (UID: "a3109e10-d592-4ff7-b6d4-82e6f07307f4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.300337 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgs25\" (UniqueName: \"kubernetes.io/projected/a3109e10-d592-4ff7-b6d4-82e6f07307f4-kube-api-access-kgs25\") on node \"crc\" DevicePath \"\"" Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.300373 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3109e10-d592-4ff7-b6d4-82e6f07307f4-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.300383 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3109e10-d592-4ff7-b6d4-82e6f07307f4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.592389 4818 generic.go:334] "Generic (PLEG): container finished" podID="a3109e10-d592-4ff7-b6d4-82e6f07307f4" containerID="b77119f3bc17040655407e94dd17c5f8e7f2acb5601246d22d29121e6cc008ee" exitCode=0 Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.592481 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzs9b" event={"ID":"a3109e10-d592-4ff7-b6d4-82e6f07307f4","Type":"ContainerDied","Data":"b77119f3bc17040655407e94dd17c5f8e7f2acb5601246d22d29121e6cc008ee"} Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.592504 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mzs9b" Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.592540 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzs9b" event={"ID":"a3109e10-d592-4ff7-b6d4-82e6f07307f4","Type":"ContainerDied","Data":"91fa872c1b5540c1f7e938c7658625c22dba370c4bb562fa14a8c4caf7b88008"} Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.592566 4818 scope.go:117] "RemoveContainer" containerID="b77119f3bc17040655407e94dd17c5f8e7f2acb5601246d22d29121e6cc008ee" Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.628524 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mzs9b"] Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.628793 4818 scope.go:117] "RemoveContainer" containerID="4d57ee0b43f124ad876eb2672ca22620e5123c8cccdea3712b9a9d001a901782" Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.642762 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mzs9b"] Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.657789 4818 scope.go:117] "RemoveContainer" containerID="168f184c196aa909e9a197b9da56cd9270501525d8bcb9a05819b34d908abfe1" Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.694442 4818 scope.go:117] "RemoveContainer" containerID="b77119f3bc17040655407e94dd17c5f8e7f2acb5601246d22d29121e6cc008ee" Nov 22 05:34:50 crc kubenswrapper[4818]: E1122 05:34:50.694910 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b77119f3bc17040655407e94dd17c5f8e7f2acb5601246d22d29121e6cc008ee\": container with ID starting with b77119f3bc17040655407e94dd17c5f8e7f2acb5601246d22d29121e6cc008ee not found: ID does not exist" containerID="b77119f3bc17040655407e94dd17c5f8e7f2acb5601246d22d29121e6cc008ee" Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.694948 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b77119f3bc17040655407e94dd17c5f8e7f2acb5601246d22d29121e6cc008ee"} err="failed to get container status \"b77119f3bc17040655407e94dd17c5f8e7f2acb5601246d22d29121e6cc008ee\": rpc error: code = NotFound desc = could not find container \"b77119f3bc17040655407e94dd17c5f8e7f2acb5601246d22d29121e6cc008ee\": container with ID starting with b77119f3bc17040655407e94dd17c5f8e7f2acb5601246d22d29121e6cc008ee not found: ID does not exist" Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.694971 4818 scope.go:117] "RemoveContainer" containerID="4d57ee0b43f124ad876eb2672ca22620e5123c8cccdea3712b9a9d001a901782" Nov 22 05:34:50 crc kubenswrapper[4818]: E1122 05:34:50.695190 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d57ee0b43f124ad876eb2672ca22620e5123c8cccdea3712b9a9d001a901782\": container with ID starting with 4d57ee0b43f124ad876eb2672ca22620e5123c8cccdea3712b9a9d001a901782 not found: ID does not exist" containerID="4d57ee0b43f124ad876eb2672ca22620e5123c8cccdea3712b9a9d001a901782" Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.695214 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d57ee0b43f124ad876eb2672ca22620e5123c8cccdea3712b9a9d001a901782"} err="failed to get container status \"4d57ee0b43f124ad876eb2672ca22620e5123c8cccdea3712b9a9d001a901782\": rpc error: code = NotFound desc = could not find container \"4d57ee0b43f124ad876eb2672ca22620e5123c8cccdea3712b9a9d001a901782\": container with ID starting with 4d57ee0b43f124ad876eb2672ca22620e5123c8cccdea3712b9a9d001a901782 not found: ID does not exist" Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.695231 4818 scope.go:117] "RemoveContainer" containerID="168f184c196aa909e9a197b9da56cd9270501525d8bcb9a05819b34d908abfe1" Nov 22 05:34:50 crc kubenswrapper[4818]: E1122 05:34:50.695547 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"168f184c196aa909e9a197b9da56cd9270501525d8bcb9a05819b34d908abfe1\": container with ID starting with 168f184c196aa909e9a197b9da56cd9270501525d8bcb9a05819b34d908abfe1 not found: ID does not exist" containerID="168f184c196aa909e9a197b9da56cd9270501525d8bcb9a05819b34d908abfe1" Nov 22 05:34:50 crc kubenswrapper[4818]: I1122 05:34:50.695574 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"168f184c196aa909e9a197b9da56cd9270501525d8bcb9a05819b34d908abfe1"} err="failed to get container status \"168f184c196aa909e9a197b9da56cd9270501525d8bcb9a05819b34d908abfe1\": rpc error: code = NotFound desc = could not find container \"168f184c196aa909e9a197b9da56cd9270501525d8bcb9a05819b34d908abfe1\": container with ID starting with 168f184c196aa909e9a197b9da56cd9270501525d8bcb9a05819b34d908abfe1 not found: ID does not exist" Nov 22 05:34:51 crc kubenswrapper[4818]: I1122 05:34:51.265325 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:34:51 crc kubenswrapper[4818]: I1122 05:34:51.265740 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:34:52 crc kubenswrapper[4818]: I1122 05:34:52.300352 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3109e10-d592-4ff7-b6d4-82e6f07307f4" path="/var/lib/kubelet/pods/a3109e10-d592-4ff7-b6d4-82e6f07307f4/volumes" Nov 22 05:35:21 crc kubenswrapper[4818]: I1122 05:35:21.264851 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:35:21 crc kubenswrapper[4818]: I1122 05:35:21.265820 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:35:21 crc kubenswrapper[4818]: I1122 05:35:21.265912 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 05:35:21 crc kubenswrapper[4818]: I1122 05:35:21.267552 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f9db7ffdf47d73a4c4ddb22a4b18596722d5f83c458d83cf3ee1c52ecf00e748"} pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 05:35:21 crc kubenswrapper[4818]: I1122 05:35:21.267730 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" containerID="cri-o://f9db7ffdf47d73a4c4ddb22a4b18596722d5f83c458d83cf3ee1c52ecf00e748" gracePeriod=600 Nov 22 05:35:21 crc kubenswrapper[4818]: I1122 05:35:21.938977 4818 generic.go:334] "Generic (PLEG): container finished" podID="0971c9ec-29ce-44eb-9875-ad4664143305" containerID="f9db7ffdf47d73a4c4ddb22a4b18596722d5f83c458d83cf3ee1c52ecf00e748" exitCode=0 Nov 22 05:35:21 crc kubenswrapper[4818]: I1122 05:35:21.939116 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerDied","Data":"f9db7ffdf47d73a4c4ddb22a4b18596722d5f83c458d83cf3ee1c52ecf00e748"} Nov 22 05:35:21 crc kubenswrapper[4818]: I1122 05:35:21.939271 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a"} Nov 22 05:35:21 crc kubenswrapper[4818]: I1122 05:35:21.939292 4818 scope.go:117] "RemoveContainer" containerID="0fbf8338b334309a18bf9b42e379e9e5fa42f77e0dff80631edeea09dfc989cc" Nov 22 05:35:27 crc kubenswrapper[4818]: I1122 05:35:27.002450 4818 generic.go:334] "Generic (PLEG): container finished" podID="0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef" containerID="0093c38d2e638f5ecb6b4c33bbf355cff805ce0aff861da9bf1041dfd69ce172" exitCode=0 Nov 22 05:35:27 crc kubenswrapper[4818]: I1122 05:35:27.002581 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" event={"ID":"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef","Type":"ContainerDied","Data":"0093c38d2e638f5ecb6b4c33bbf355cff805ce0aff861da9bf1041dfd69ce172"} Nov 22 05:35:28 crc kubenswrapper[4818]: I1122 05:35:28.519990 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:35:28 crc kubenswrapper[4818]: I1122 05:35:28.644641 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ovn-combined-ca-bundle\") pod \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " Nov 22 05:35:28 crc kubenswrapper[4818]: I1122 05:35:28.644703 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ovncontroller-config-0\") pod \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " Nov 22 05:35:28 crc kubenswrapper[4818]: I1122 05:35:28.644814 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-inventory\") pod \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " Nov 22 05:35:28 crc kubenswrapper[4818]: I1122 05:35:28.644869 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ssh-key\") pod \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " Nov 22 05:35:28 crc kubenswrapper[4818]: I1122 05:35:28.644901 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znthj\" (UniqueName: \"kubernetes.io/projected/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-kube-api-access-znthj\") pod \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " Nov 22 05:35:28 crc kubenswrapper[4818]: I1122 05:35:28.644940 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ceph\") pod \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\" (UID: \"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef\") " Nov 22 05:35:28 crc kubenswrapper[4818]: I1122 05:35:28.651704 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-kube-api-access-znthj" (OuterVolumeSpecName: "kube-api-access-znthj") pod "0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef" (UID: "0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef"). InnerVolumeSpecName "kube-api-access-znthj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:35:28 crc kubenswrapper[4818]: I1122 05:35:28.653303 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef" (UID: "0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:35:28 crc kubenswrapper[4818]: I1122 05:35:28.653425 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ceph" (OuterVolumeSpecName: "ceph") pod "0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef" (UID: "0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:35:28 crc kubenswrapper[4818]: I1122 05:35:28.674099 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef" (UID: "0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:35:28 crc kubenswrapper[4818]: I1122 05:35:28.677477 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef" (UID: "0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:35:28 crc kubenswrapper[4818]: I1122 05:35:28.695285 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-inventory" (OuterVolumeSpecName: "inventory") pod "0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef" (UID: "0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:35:28 crc kubenswrapper[4818]: I1122 05:35:28.746530 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:35:28 crc kubenswrapper[4818]: I1122 05:35:28.746562 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znthj\" (UniqueName: \"kubernetes.io/projected/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-kube-api-access-znthj\") on node \"crc\" DevicePath \"\"" Nov 22 05:35:28 crc kubenswrapper[4818]: I1122 05:35:28.746575 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:35:28 crc kubenswrapper[4818]: I1122 05:35:28.746585 4818 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:35:28 crc kubenswrapper[4818]: I1122 05:35:28.746595 4818 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 22 05:35:28 crc kubenswrapper[4818]: I1122 05:35:28.746604 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.028682 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" event={"ID":"0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef","Type":"ContainerDied","Data":"d93db8c9deaba62e859bb688db5086926d99180e181bb858471b0d4c420e2f7b"} Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.028785 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d93db8c9deaba62e859bb688db5086926d99180e181bb858471b0d4c420e2f7b" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.028735 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k2rlm" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.163297 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z"] Nov 22 05:35:29 crc kubenswrapper[4818]: E1122 05:35:29.163721 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3109e10-d592-4ff7-b6d4-82e6f07307f4" containerName="registry-server" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.163743 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3109e10-d592-4ff7-b6d4-82e6f07307f4" containerName="registry-server" Nov 22 05:35:29 crc kubenswrapper[4818]: E1122 05:35:29.163765 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3109e10-d592-4ff7-b6d4-82e6f07307f4" containerName="extract-utilities" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.163773 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3109e10-d592-4ff7-b6d4-82e6f07307f4" containerName="extract-utilities" Nov 22 05:35:29 crc kubenswrapper[4818]: E1122 05:35:29.163784 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3109e10-d592-4ff7-b6d4-82e6f07307f4" containerName="extract-content" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.163793 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3109e10-d592-4ff7-b6d4-82e6f07307f4" containerName="extract-content" Nov 22 05:35:29 crc kubenswrapper[4818]: E1122 05:35:29.163813 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.163821 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.164046 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3109e10-d592-4ff7-b6d4-82e6f07307f4" containerName="registry-server" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.164085 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.165055 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.168019 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.170399 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.170773 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.171504 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.171899 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.172551 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.172912 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.181762 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z"] Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.255529 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.255590 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.255627 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.255658 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.255684 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvvv5\" (UniqueName: \"kubernetes.io/projected/e80824ae-495b-475f-a266-44879bb04e25-kube-api-access-pvvv5\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.255983 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.256057 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.357994 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.358059 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.358149 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.358180 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.358209 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.358248 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.358287 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvvv5\" (UniqueName: \"kubernetes.io/projected/e80824ae-495b-475f-a266-44879bb04e25-kube-api-access-pvvv5\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.363083 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.363615 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.363817 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.364533 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.364707 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.365307 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.379387 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvvv5\" (UniqueName: \"kubernetes.io/projected/e80824ae-495b-475f-a266-44879bb04e25-kube-api-access-pvvv5\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:29 crc kubenswrapper[4818]: I1122 05:35:29.495357 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:35:30 crc kubenswrapper[4818]: I1122 05:35:30.085634 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z"] Nov 22 05:35:30 crc kubenswrapper[4818]: W1122 05:35:30.092545 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode80824ae_495b_475f_a266_44879bb04e25.slice/crio-d0a58fd37c016a315378071bc533afd56807fef864d1b8ee65b86be148340131 WatchSource:0}: Error finding container d0a58fd37c016a315378071bc533afd56807fef864d1b8ee65b86be148340131: Status 404 returned error can't find the container with id d0a58fd37c016a315378071bc533afd56807fef864d1b8ee65b86be148340131 Nov 22 05:35:31 crc kubenswrapper[4818]: I1122 05:35:31.052713 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" event={"ID":"e80824ae-495b-475f-a266-44879bb04e25","Type":"ContainerStarted","Data":"e8b53832dcfc05140b55b880a8b9de12541df7efb57b9f66eb4b2e86a217c16c"} Nov 22 05:35:31 crc kubenswrapper[4818]: I1122 05:35:31.053580 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" event={"ID":"e80824ae-495b-475f-a266-44879bb04e25","Type":"ContainerStarted","Data":"d0a58fd37c016a315378071bc533afd56807fef864d1b8ee65b86be148340131"} Nov 22 05:36:36 crc kubenswrapper[4818]: I1122 05:36:36.732712 4818 generic.go:334] "Generic (PLEG): container finished" podID="e80824ae-495b-475f-a266-44879bb04e25" containerID="e8b53832dcfc05140b55b880a8b9de12541df7efb57b9f66eb4b2e86a217c16c" exitCode=0 Nov 22 05:36:36 crc kubenswrapper[4818]: I1122 05:36:36.733368 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" event={"ID":"e80824ae-495b-475f-a266-44879bb04e25","Type":"ContainerDied","Data":"e8b53832dcfc05140b55b880a8b9de12541df7efb57b9f66eb4b2e86a217c16c"} Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.110369 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.223931 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-nova-metadata-neutron-config-0\") pod \"e80824ae-495b-475f-a266-44879bb04e25\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.224051 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-inventory\") pod \"e80824ae-495b-475f-a266-44879bb04e25\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.224095 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-ssh-key\") pod \"e80824ae-495b-475f-a266-44879bb04e25\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.224121 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-neutron-metadata-combined-ca-bundle\") pod \"e80824ae-495b-475f-a266-44879bb04e25\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.224215 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvvv5\" (UniqueName: \"kubernetes.io/projected/e80824ae-495b-475f-a266-44879bb04e25-kube-api-access-pvvv5\") pod \"e80824ae-495b-475f-a266-44879bb04e25\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.224295 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-neutron-ovn-metadata-agent-neutron-config-0\") pod \"e80824ae-495b-475f-a266-44879bb04e25\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.224331 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-ceph\") pod \"e80824ae-495b-475f-a266-44879bb04e25\" (UID: \"e80824ae-495b-475f-a266-44879bb04e25\") " Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.229054 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "e80824ae-495b-475f-a266-44879bb04e25" (UID: "e80824ae-495b-475f-a266-44879bb04e25"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.229087 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e80824ae-495b-475f-a266-44879bb04e25-kube-api-access-pvvv5" (OuterVolumeSpecName: "kube-api-access-pvvv5") pod "e80824ae-495b-475f-a266-44879bb04e25" (UID: "e80824ae-495b-475f-a266-44879bb04e25"). InnerVolumeSpecName "kube-api-access-pvvv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.229790 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-ceph" (OuterVolumeSpecName: "ceph") pod "e80824ae-495b-475f-a266-44879bb04e25" (UID: "e80824ae-495b-475f-a266-44879bb04e25"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.251207 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e80824ae-495b-475f-a266-44879bb04e25" (UID: "e80824ae-495b-475f-a266-44879bb04e25"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.254287 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "e80824ae-495b-475f-a266-44879bb04e25" (UID: "e80824ae-495b-475f-a266-44879bb04e25"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.258094 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-inventory" (OuterVolumeSpecName: "inventory") pod "e80824ae-495b-475f-a266-44879bb04e25" (UID: "e80824ae-495b-475f-a266-44879bb04e25"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.263396 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "e80824ae-495b-475f-a266-44879bb04e25" (UID: "e80824ae-495b-475f-a266-44879bb04e25"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.327616 4818 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.327662 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.327681 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.327699 4818 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.327718 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvvv5\" (UniqueName: \"kubernetes.io/projected/e80824ae-495b-475f-a266-44879bb04e25-kube-api-access-pvvv5\") on node \"crc\" DevicePath \"\"" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.327735 4818 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.327753 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e80824ae-495b-475f-a266-44879bb04e25-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.756926 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" event={"ID":"e80824ae-495b-475f-a266-44879bb04e25","Type":"ContainerDied","Data":"d0a58fd37c016a315378071bc533afd56807fef864d1b8ee65b86be148340131"} Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.757242 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0a58fd37c016a315378071bc533afd56807fef864d1b8ee65b86be148340131" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.757010 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.947743 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb"] Nov 22 05:36:38 crc kubenswrapper[4818]: E1122 05:36:38.948132 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e80824ae-495b-475f-a266-44879bb04e25" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.948148 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="e80824ae-495b-475f-a266-44879bb04e25" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.948352 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="e80824ae-495b-475f-a266-44879bb04e25" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.948987 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.954731 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.954919 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.955047 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.955196 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.955412 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.955656 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:36:38 crc kubenswrapper[4818]: I1122 05:36:38.964707 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb"] Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.042133 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-595lb\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.042206 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-595lb\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.042281 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-595lb\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.042316 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-595lb\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.042354 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-595lb\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.042374 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzlv8\" (UniqueName: \"kubernetes.io/projected/806c9298-1aad-4e84-a80a-b2078f548d03-kube-api-access-lzlv8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-595lb\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.144357 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-595lb\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.144482 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-595lb\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.144553 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-595lb\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.144602 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-595lb\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.144652 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-595lb\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.144679 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzlv8\" (UniqueName: \"kubernetes.io/projected/806c9298-1aad-4e84-a80a-b2078f548d03-kube-api-access-lzlv8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-595lb\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.149426 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-595lb\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.149562 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-595lb\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.149625 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-595lb\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.149983 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-595lb\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.150480 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-595lb\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.163353 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzlv8\" (UniqueName: \"kubernetes.io/projected/806c9298-1aad-4e84-a80a-b2078f548d03-kube-api-access-lzlv8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-595lb\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.265076 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.650445 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb"] Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.654204 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 05:36:39 crc kubenswrapper[4818]: I1122 05:36:39.766375 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" event={"ID":"806c9298-1aad-4e84-a80a-b2078f548d03","Type":"ContainerStarted","Data":"5ef0e14dae48201f24c7a94c3957558ad2320fc04d32e643522a9ea85874e0de"} Nov 22 05:36:40 crc kubenswrapper[4818]: I1122 05:36:40.779612 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" event={"ID":"806c9298-1aad-4e84-a80a-b2078f548d03","Type":"ContainerStarted","Data":"713152ea6d7bc3096a4fc0747a6daee2bcbfb4708d61eeaa6ff42dc79c5f1b0d"} Nov 22 05:36:40 crc kubenswrapper[4818]: I1122 05:36:40.814728 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" podStartSLOduration=2.227528188 podStartE2EDuration="2.814699664s" podCreationTimestamp="2025-11-22 05:36:38 +0000 UTC" firstStartedPulling="2025-11-22 05:36:39.654004926 +0000 UTC m=+2952.228421453" lastFinishedPulling="2025-11-22 05:36:40.241176412 +0000 UTC m=+2952.815592929" observedRunningTime="2025-11-22 05:36:40.803814943 +0000 UTC m=+2953.378231470" watchObservedRunningTime="2025-11-22 05:36:40.814699664 +0000 UTC m=+2953.389116221" Nov 22 05:37:11 crc kubenswrapper[4818]: I1122 05:37:11.600200 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rnm24"] Nov 22 05:37:11 crc kubenswrapper[4818]: I1122 05:37:11.614017 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rnm24"] Nov 22 05:37:11 crc kubenswrapper[4818]: I1122 05:37:11.614340 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rnm24" Nov 22 05:37:11 crc kubenswrapper[4818]: I1122 05:37:11.673031 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b7bb1f-3d8c-41bb-8933-a9904a04b635-catalog-content\") pod \"community-operators-rnm24\" (UID: \"59b7bb1f-3d8c-41bb-8933-a9904a04b635\") " pod="openshift-marketplace/community-operators-rnm24" Nov 22 05:37:11 crc kubenswrapper[4818]: I1122 05:37:11.673082 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh6js\" (UniqueName: \"kubernetes.io/projected/59b7bb1f-3d8c-41bb-8933-a9904a04b635-kube-api-access-wh6js\") pod \"community-operators-rnm24\" (UID: \"59b7bb1f-3d8c-41bb-8933-a9904a04b635\") " pod="openshift-marketplace/community-operators-rnm24" Nov 22 05:37:11 crc kubenswrapper[4818]: I1122 05:37:11.673138 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b7bb1f-3d8c-41bb-8933-a9904a04b635-utilities\") pod \"community-operators-rnm24\" (UID: \"59b7bb1f-3d8c-41bb-8933-a9904a04b635\") " pod="openshift-marketplace/community-operators-rnm24" Nov 22 05:37:11 crc kubenswrapper[4818]: I1122 05:37:11.775179 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b7bb1f-3d8c-41bb-8933-a9904a04b635-utilities\") pod \"community-operators-rnm24\" (UID: \"59b7bb1f-3d8c-41bb-8933-a9904a04b635\") " pod="openshift-marketplace/community-operators-rnm24" Nov 22 05:37:11 crc kubenswrapper[4818]: I1122 05:37:11.775383 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b7bb1f-3d8c-41bb-8933-a9904a04b635-catalog-content\") pod \"community-operators-rnm24\" (UID: \"59b7bb1f-3d8c-41bb-8933-a9904a04b635\") " pod="openshift-marketplace/community-operators-rnm24" Nov 22 05:37:11 crc kubenswrapper[4818]: I1122 05:37:11.775407 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh6js\" (UniqueName: \"kubernetes.io/projected/59b7bb1f-3d8c-41bb-8933-a9904a04b635-kube-api-access-wh6js\") pod \"community-operators-rnm24\" (UID: \"59b7bb1f-3d8c-41bb-8933-a9904a04b635\") " pod="openshift-marketplace/community-operators-rnm24" Nov 22 05:37:11 crc kubenswrapper[4818]: I1122 05:37:11.776054 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b7bb1f-3d8c-41bb-8933-a9904a04b635-catalog-content\") pod \"community-operators-rnm24\" (UID: \"59b7bb1f-3d8c-41bb-8933-a9904a04b635\") " pod="openshift-marketplace/community-operators-rnm24" Nov 22 05:37:11 crc kubenswrapper[4818]: I1122 05:37:11.776067 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b7bb1f-3d8c-41bb-8933-a9904a04b635-utilities\") pod \"community-operators-rnm24\" (UID: \"59b7bb1f-3d8c-41bb-8933-a9904a04b635\") " pod="openshift-marketplace/community-operators-rnm24" Nov 22 05:37:11 crc kubenswrapper[4818]: I1122 05:37:11.794956 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh6js\" (UniqueName: \"kubernetes.io/projected/59b7bb1f-3d8c-41bb-8933-a9904a04b635-kube-api-access-wh6js\") pod \"community-operators-rnm24\" (UID: \"59b7bb1f-3d8c-41bb-8933-a9904a04b635\") " pod="openshift-marketplace/community-operators-rnm24" Nov 22 05:37:11 crc kubenswrapper[4818]: I1122 05:37:11.939806 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rnm24" Nov 22 05:37:12 crc kubenswrapper[4818]: I1122 05:37:12.481462 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rnm24"] Nov 22 05:37:12 crc kubenswrapper[4818]: W1122 05:37:12.488938 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59b7bb1f_3d8c_41bb_8933_a9904a04b635.slice/crio-e2b954702bacffe7407ef5b5e5676ed0799a17439d699bee1fe4977765e46bb5 WatchSource:0}: Error finding container e2b954702bacffe7407ef5b5e5676ed0799a17439d699bee1fe4977765e46bb5: Status 404 returned error can't find the container with id e2b954702bacffe7407ef5b5e5676ed0799a17439d699bee1fe4977765e46bb5 Nov 22 05:37:13 crc kubenswrapper[4818]: I1122 05:37:13.181911 4818 generic.go:334] "Generic (PLEG): container finished" podID="59b7bb1f-3d8c-41bb-8933-a9904a04b635" containerID="28c63feddbd01479f483c8f8746885b1fc92eadaef85369a8090d4d9332d41f2" exitCode=0 Nov 22 05:37:13 crc kubenswrapper[4818]: I1122 05:37:13.181981 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnm24" event={"ID":"59b7bb1f-3d8c-41bb-8933-a9904a04b635","Type":"ContainerDied","Data":"28c63feddbd01479f483c8f8746885b1fc92eadaef85369a8090d4d9332d41f2"} Nov 22 05:37:13 crc kubenswrapper[4818]: I1122 05:37:13.182021 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnm24" event={"ID":"59b7bb1f-3d8c-41bb-8933-a9904a04b635","Type":"ContainerStarted","Data":"e2b954702bacffe7407ef5b5e5676ed0799a17439d699bee1fe4977765e46bb5"} Nov 22 05:37:15 crc kubenswrapper[4818]: I1122 05:37:15.205152 4818 generic.go:334] "Generic (PLEG): container finished" podID="59b7bb1f-3d8c-41bb-8933-a9904a04b635" containerID="a9fbfff1e87d22766f62b7765f6d6244e8c48e7827988bc9c5b62c39878274fb" exitCode=0 Nov 22 05:37:15 crc kubenswrapper[4818]: I1122 05:37:15.205333 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnm24" event={"ID":"59b7bb1f-3d8c-41bb-8933-a9904a04b635","Type":"ContainerDied","Data":"a9fbfff1e87d22766f62b7765f6d6244e8c48e7827988bc9c5b62c39878274fb"} Nov 22 05:37:17 crc kubenswrapper[4818]: I1122 05:37:17.227758 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnm24" event={"ID":"59b7bb1f-3d8c-41bb-8933-a9904a04b635","Type":"ContainerStarted","Data":"ae239442a09adf3d7399cac0a2a65b3355419d11e998d8afbf3f8472dd2a718b"} Nov 22 05:37:17 crc kubenswrapper[4818]: I1122 05:37:17.255183 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rnm24" podStartSLOduration=3.37981991 podStartE2EDuration="6.255156093s" podCreationTimestamp="2025-11-22 05:37:11 +0000 UTC" firstStartedPulling="2025-11-22 05:37:13.184884828 +0000 UTC m=+2985.759301395" lastFinishedPulling="2025-11-22 05:37:16.060221041 +0000 UTC m=+2988.634637578" observedRunningTime="2025-11-22 05:37:17.24791013 +0000 UTC m=+2989.822326687" watchObservedRunningTime="2025-11-22 05:37:17.255156093 +0000 UTC m=+2989.829572660" Nov 22 05:37:21 crc kubenswrapper[4818]: I1122 05:37:21.264543 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:37:21 crc kubenswrapper[4818]: I1122 05:37:21.265338 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:37:21 crc kubenswrapper[4818]: I1122 05:37:21.940107 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rnm24" Nov 22 05:37:21 crc kubenswrapper[4818]: I1122 05:37:21.940161 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rnm24" Nov 22 05:37:22 crc kubenswrapper[4818]: I1122 05:37:22.008024 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rnm24" Nov 22 05:37:22 crc kubenswrapper[4818]: I1122 05:37:22.363268 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rnm24" Nov 22 05:37:22 crc kubenswrapper[4818]: I1122 05:37:22.429575 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rnm24"] Nov 22 05:37:24 crc kubenswrapper[4818]: I1122 05:37:24.295760 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rnm24" podUID="59b7bb1f-3d8c-41bb-8933-a9904a04b635" containerName="registry-server" containerID="cri-o://ae239442a09adf3d7399cac0a2a65b3355419d11e998d8afbf3f8472dd2a718b" gracePeriod=2 Nov 22 05:37:24 crc kubenswrapper[4818]: I1122 05:37:24.760015 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rnm24" Nov 22 05:37:24 crc kubenswrapper[4818]: I1122 05:37:24.945191 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b7bb1f-3d8c-41bb-8933-a9904a04b635-utilities\") pod \"59b7bb1f-3d8c-41bb-8933-a9904a04b635\" (UID: \"59b7bb1f-3d8c-41bb-8933-a9904a04b635\") " Nov 22 05:37:24 crc kubenswrapper[4818]: I1122 05:37:24.945420 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wh6js\" (UniqueName: \"kubernetes.io/projected/59b7bb1f-3d8c-41bb-8933-a9904a04b635-kube-api-access-wh6js\") pod \"59b7bb1f-3d8c-41bb-8933-a9904a04b635\" (UID: \"59b7bb1f-3d8c-41bb-8933-a9904a04b635\") " Nov 22 05:37:24 crc kubenswrapper[4818]: I1122 05:37:24.945720 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b7bb1f-3d8c-41bb-8933-a9904a04b635-catalog-content\") pod \"59b7bb1f-3d8c-41bb-8933-a9904a04b635\" (UID: \"59b7bb1f-3d8c-41bb-8933-a9904a04b635\") " Nov 22 05:37:24 crc kubenswrapper[4818]: I1122 05:37:24.946049 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59b7bb1f-3d8c-41bb-8933-a9904a04b635-utilities" (OuterVolumeSpecName: "utilities") pod "59b7bb1f-3d8c-41bb-8933-a9904a04b635" (UID: "59b7bb1f-3d8c-41bb-8933-a9904a04b635"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:37:24 crc kubenswrapper[4818]: I1122 05:37:24.946235 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b7bb1f-3d8c-41bb-8933-a9904a04b635-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 05:37:24 crc kubenswrapper[4818]: I1122 05:37:24.950859 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59b7bb1f-3d8c-41bb-8933-a9904a04b635-kube-api-access-wh6js" (OuterVolumeSpecName: "kube-api-access-wh6js") pod "59b7bb1f-3d8c-41bb-8933-a9904a04b635" (UID: "59b7bb1f-3d8c-41bb-8933-a9904a04b635"). InnerVolumeSpecName "kube-api-access-wh6js". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:37:25 crc kubenswrapper[4818]: I1122 05:37:25.047970 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wh6js\" (UniqueName: \"kubernetes.io/projected/59b7bb1f-3d8c-41bb-8933-a9904a04b635-kube-api-access-wh6js\") on node \"crc\" DevicePath \"\"" Nov 22 05:37:25 crc kubenswrapper[4818]: I1122 05:37:25.313744 4818 generic.go:334] "Generic (PLEG): container finished" podID="59b7bb1f-3d8c-41bb-8933-a9904a04b635" containerID="ae239442a09adf3d7399cac0a2a65b3355419d11e998d8afbf3f8472dd2a718b" exitCode=0 Nov 22 05:37:25 crc kubenswrapper[4818]: I1122 05:37:25.313822 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnm24" event={"ID":"59b7bb1f-3d8c-41bb-8933-a9904a04b635","Type":"ContainerDied","Data":"ae239442a09adf3d7399cac0a2a65b3355419d11e998d8afbf3f8472dd2a718b"} Nov 22 05:37:25 crc kubenswrapper[4818]: I1122 05:37:25.313869 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnm24" event={"ID":"59b7bb1f-3d8c-41bb-8933-a9904a04b635","Type":"ContainerDied","Data":"e2b954702bacffe7407ef5b5e5676ed0799a17439d699bee1fe4977765e46bb5"} Nov 22 05:37:25 crc kubenswrapper[4818]: I1122 05:37:25.313910 4818 scope.go:117] "RemoveContainer" containerID="ae239442a09adf3d7399cac0a2a65b3355419d11e998d8afbf3f8472dd2a718b" Nov 22 05:37:25 crc kubenswrapper[4818]: I1122 05:37:25.314196 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rnm24" Nov 22 05:37:25 crc kubenswrapper[4818]: I1122 05:37:25.348852 4818 scope.go:117] "RemoveContainer" containerID="a9fbfff1e87d22766f62b7765f6d6244e8c48e7827988bc9c5b62c39878274fb" Nov 22 05:37:25 crc kubenswrapper[4818]: I1122 05:37:25.377830 4818 scope.go:117] "RemoveContainer" containerID="28c63feddbd01479f483c8f8746885b1fc92eadaef85369a8090d4d9332d41f2" Nov 22 05:37:25 crc kubenswrapper[4818]: I1122 05:37:25.443664 4818 scope.go:117] "RemoveContainer" containerID="ae239442a09adf3d7399cac0a2a65b3355419d11e998d8afbf3f8472dd2a718b" Nov 22 05:37:25 crc kubenswrapper[4818]: E1122 05:37:25.444878 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae239442a09adf3d7399cac0a2a65b3355419d11e998d8afbf3f8472dd2a718b\": container with ID starting with ae239442a09adf3d7399cac0a2a65b3355419d11e998d8afbf3f8472dd2a718b not found: ID does not exist" containerID="ae239442a09adf3d7399cac0a2a65b3355419d11e998d8afbf3f8472dd2a718b" Nov 22 05:37:25 crc kubenswrapper[4818]: I1122 05:37:25.444964 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae239442a09adf3d7399cac0a2a65b3355419d11e998d8afbf3f8472dd2a718b"} err="failed to get container status \"ae239442a09adf3d7399cac0a2a65b3355419d11e998d8afbf3f8472dd2a718b\": rpc error: code = NotFound desc = could not find container \"ae239442a09adf3d7399cac0a2a65b3355419d11e998d8afbf3f8472dd2a718b\": container with ID starting with ae239442a09adf3d7399cac0a2a65b3355419d11e998d8afbf3f8472dd2a718b not found: ID does not exist" Nov 22 05:37:25 crc kubenswrapper[4818]: I1122 05:37:25.445019 4818 scope.go:117] "RemoveContainer" containerID="a9fbfff1e87d22766f62b7765f6d6244e8c48e7827988bc9c5b62c39878274fb" Nov 22 05:37:25 crc kubenswrapper[4818]: E1122 05:37:25.445704 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9fbfff1e87d22766f62b7765f6d6244e8c48e7827988bc9c5b62c39878274fb\": container with ID starting with a9fbfff1e87d22766f62b7765f6d6244e8c48e7827988bc9c5b62c39878274fb not found: ID does not exist" containerID="a9fbfff1e87d22766f62b7765f6d6244e8c48e7827988bc9c5b62c39878274fb" Nov 22 05:37:25 crc kubenswrapper[4818]: I1122 05:37:25.445739 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9fbfff1e87d22766f62b7765f6d6244e8c48e7827988bc9c5b62c39878274fb"} err="failed to get container status \"a9fbfff1e87d22766f62b7765f6d6244e8c48e7827988bc9c5b62c39878274fb\": rpc error: code = NotFound desc = could not find container \"a9fbfff1e87d22766f62b7765f6d6244e8c48e7827988bc9c5b62c39878274fb\": container with ID starting with a9fbfff1e87d22766f62b7765f6d6244e8c48e7827988bc9c5b62c39878274fb not found: ID does not exist" Nov 22 05:37:25 crc kubenswrapper[4818]: I1122 05:37:25.445762 4818 scope.go:117] "RemoveContainer" containerID="28c63feddbd01479f483c8f8746885b1fc92eadaef85369a8090d4d9332d41f2" Nov 22 05:37:25 crc kubenswrapper[4818]: E1122 05:37:25.446062 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28c63feddbd01479f483c8f8746885b1fc92eadaef85369a8090d4d9332d41f2\": container with ID starting with 28c63feddbd01479f483c8f8746885b1fc92eadaef85369a8090d4d9332d41f2 not found: ID does not exist" containerID="28c63feddbd01479f483c8f8746885b1fc92eadaef85369a8090d4d9332d41f2" Nov 22 05:37:25 crc kubenswrapper[4818]: I1122 05:37:25.446105 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28c63feddbd01479f483c8f8746885b1fc92eadaef85369a8090d4d9332d41f2"} err="failed to get container status \"28c63feddbd01479f483c8f8746885b1fc92eadaef85369a8090d4d9332d41f2\": rpc error: code = NotFound desc = could not find container \"28c63feddbd01479f483c8f8746885b1fc92eadaef85369a8090d4d9332d41f2\": container with ID starting with 28c63feddbd01479f483c8f8746885b1fc92eadaef85369a8090d4d9332d41f2 not found: ID does not exist" Nov 22 05:37:25 crc kubenswrapper[4818]: I1122 05:37:25.766913 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59b7bb1f-3d8c-41bb-8933-a9904a04b635-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59b7bb1f-3d8c-41bb-8933-a9904a04b635" (UID: "59b7bb1f-3d8c-41bb-8933-a9904a04b635"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:37:25 crc kubenswrapper[4818]: I1122 05:37:25.869996 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b7bb1f-3d8c-41bb-8933-a9904a04b635-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 05:37:25 crc kubenswrapper[4818]: I1122 05:37:25.961295 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rnm24"] Nov 22 05:37:25 crc kubenswrapper[4818]: I1122 05:37:25.971072 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rnm24"] Nov 22 05:37:26 crc kubenswrapper[4818]: I1122 05:37:26.304314 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59b7bb1f-3d8c-41bb-8933-a9904a04b635" path="/var/lib/kubelet/pods/59b7bb1f-3d8c-41bb-8933-a9904a04b635/volumes" Nov 22 05:37:51 crc kubenswrapper[4818]: I1122 05:37:51.264743 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:37:51 crc kubenswrapper[4818]: I1122 05:37:51.265463 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:38:21 crc kubenswrapper[4818]: I1122 05:38:21.264735 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:38:21 crc kubenswrapper[4818]: I1122 05:38:21.265349 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:38:21 crc kubenswrapper[4818]: I1122 05:38:21.265413 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 05:38:21 crc kubenswrapper[4818]: I1122 05:38:21.266679 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a"} pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 05:38:21 crc kubenswrapper[4818]: I1122 05:38:21.266832 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" containerID="cri-o://297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" gracePeriod=600 Nov 22 05:38:21 crc kubenswrapper[4818]: E1122 05:38:21.399155 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:38:21 crc kubenswrapper[4818]: I1122 05:38:21.947537 4818 generic.go:334] "Generic (PLEG): container finished" podID="0971c9ec-29ce-44eb-9875-ad4664143305" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" exitCode=0 Nov 22 05:38:21 crc kubenswrapper[4818]: I1122 05:38:21.947632 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerDied","Data":"297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a"} Nov 22 05:38:21 crc kubenswrapper[4818]: I1122 05:38:21.947921 4818 scope.go:117] "RemoveContainer" containerID="f9db7ffdf47d73a4c4ddb22a4b18596722d5f83c458d83cf3ee1c52ecf00e748" Nov 22 05:38:21 crc kubenswrapper[4818]: I1122 05:38:21.948586 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:38:21 crc kubenswrapper[4818]: E1122 05:38:21.948873 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:38:32 crc kubenswrapper[4818]: I1122 05:38:32.290993 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:38:32 crc kubenswrapper[4818]: E1122 05:38:32.292885 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:38:46 crc kubenswrapper[4818]: I1122 05:38:46.290921 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:38:46 crc kubenswrapper[4818]: E1122 05:38:46.291700 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:38:59 crc kubenswrapper[4818]: I1122 05:38:59.291493 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:38:59 crc kubenswrapper[4818]: E1122 05:38:59.292614 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:39:12 crc kubenswrapper[4818]: I1122 05:39:12.291360 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:39:12 crc kubenswrapper[4818]: E1122 05:39:12.292388 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:39:23 crc kubenswrapper[4818]: I1122 05:39:23.292071 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:39:23 crc kubenswrapper[4818]: E1122 05:39:23.293127 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:39:36 crc kubenswrapper[4818]: I1122 05:39:36.291407 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:39:36 crc kubenswrapper[4818]: E1122 05:39:36.291995 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:39:47 crc kubenswrapper[4818]: I1122 05:39:47.292577 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:39:47 crc kubenswrapper[4818]: E1122 05:39:47.293620 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:39:59 crc kubenswrapper[4818]: I1122 05:39:59.292130 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:39:59 crc kubenswrapper[4818]: E1122 05:39:59.293204 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:40:13 crc kubenswrapper[4818]: I1122 05:40:13.291473 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:40:13 crc kubenswrapper[4818]: E1122 05:40:13.292521 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:40:25 crc kubenswrapper[4818]: I1122 05:40:25.292016 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:40:25 crc kubenswrapper[4818]: E1122 05:40:25.293112 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:40:36 crc kubenswrapper[4818]: I1122 05:40:36.292827 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:40:36 crc kubenswrapper[4818]: E1122 05:40:36.294415 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:40:50 crc kubenswrapper[4818]: I1122 05:40:50.291207 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:40:50 crc kubenswrapper[4818]: E1122 05:40:50.292080 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:41:03 crc kubenswrapper[4818]: I1122 05:41:03.292003 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:41:03 crc kubenswrapper[4818]: E1122 05:41:03.293061 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:41:18 crc kubenswrapper[4818]: I1122 05:41:18.303347 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:41:18 crc kubenswrapper[4818]: E1122 05:41:18.304741 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:41:33 crc kubenswrapper[4818]: I1122 05:41:33.292359 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:41:33 crc kubenswrapper[4818]: E1122 05:41:33.293587 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:41:35 crc kubenswrapper[4818]: I1122 05:41:35.136173 4818 generic.go:334] "Generic (PLEG): container finished" podID="806c9298-1aad-4e84-a80a-b2078f548d03" containerID="713152ea6d7bc3096a4fc0747a6daee2bcbfb4708d61eeaa6ff42dc79c5f1b0d" exitCode=0 Nov 22 05:41:35 crc kubenswrapper[4818]: I1122 05:41:35.136487 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" event={"ID":"806c9298-1aad-4e84-a80a-b2078f548d03","Type":"ContainerDied","Data":"713152ea6d7bc3096a4fc0747a6daee2bcbfb4708d61eeaa6ff42dc79c5f1b0d"} Nov 22 05:41:36 crc kubenswrapper[4818]: I1122 05:41:36.502240 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:41:36 crc kubenswrapper[4818]: I1122 05:41:36.610604 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-libvirt-combined-ca-bundle\") pod \"806c9298-1aad-4e84-a80a-b2078f548d03\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " Nov 22 05:41:36 crc kubenswrapper[4818]: I1122 05:41:36.610673 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-ssh-key\") pod \"806c9298-1aad-4e84-a80a-b2078f548d03\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " Nov 22 05:41:36 crc kubenswrapper[4818]: I1122 05:41:36.610729 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzlv8\" (UniqueName: \"kubernetes.io/projected/806c9298-1aad-4e84-a80a-b2078f548d03-kube-api-access-lzlv8\") pod \"806c9298-1aad-4e84-a80a-b2078f548d03\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " Nov 22 05:41:36 crc kubenswrapper[4818]: I1122 05:41:36.610828 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-inventory\") pod \"806c9298-1aad-4e84-a80a-b2078f548d03\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " Nov 22 05:41:36 crc kubenswrapper[4818]: I1122 05:41:36.610880 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-ceph\") pod \"806c9298-1aad-4e84-a80a-b2078f548d03\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " Nov 22 05:41:36 crc kubenswrapper[4818]: I1122 05:41:36.610919 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-libvirt-secret-0\") pod \"806c9298-1aad-4e84-a80a-b2078f548d03\" (UID: \"806c9298-1aad-4e84-a80a-b2078f548d03\") " Nov 22 05:41:36 crc kubenswrapper[4818]: I1122 05:41:36.616994 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/806c9298-1aad-4e84-a80a-b2078f548d03-kube-api-access-lzlv8" (OuterVolumeSpecName: "kube-api-access-lzlv8") pod "806c9298-1aad-4e84-a80a-b2078f548d03" (UID: "806c9298-1aad-4e84-a80a-b2078f548d03"). InnerVolumeSpecName "kube-api-access-lzlv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:41:36 crc kubenswrapper[4818]: I1122 05:41:36.617393 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "806c9298-1aad-4e84-a80a-b2078f548d03" (UID: "806c9298-1aad-4e84-a80a-b2078f548d03"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:41:36 crc kubenswrapper[4818]: I1122 05:41:36.617759 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-ceph" (OuterVolumeSpecName: "ceph") pod "806c9298-1aad-4e84-a80a-b2078f548d03" (UID: "806c9298-1aad-4e84-a80a-b2078f548d03"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:41:36 crc kubenswrapper[4818]: I1122 05:41:36.639388 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "806c9298-1aad-4e84-a80a-b2078f548d03" (UID: "806c9298-1aad-4e84-a80a-b2078f548d03"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:41:36 crc kubenswrapper[4818]: I1122 05:41:36.646657 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "806c9298-1aad-4e84-a80a-b2078f548d03" (UID: "806c9298-1aad-4e84-a80a-b2078f548d03"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:41:36 crc kubenswrapper[4818]: I1122 05:41:36.659976 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-inventory" (OuterVolumeSpecName: "inventory") pod "806c9298-1aad-4e84-a80a-b2078f548d03" (UID: "806c9298-1aad-4e84-a80a-b2078f548d03"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:41:36 crc kubenswrapper[4818]: I1122 05:41:36.712691 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:41:36 crc kubenswrapper[4818]: I1122 05:41:36.712713 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:41:36 crc kubenswrapper[4818]: I1122 05:41:36.712724 4818 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 22 05:41:36 crc kubenswrapper[4818]: I1122 05:41:36.712736 4818 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:41:36 crc kubenswrapper[4818]: I1122 05:41:36.712745 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/806c9298-1aad-4e84-a80a-b2078f548d03-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:41:36 crc kubenswrapper[4818]: I1122 05:41:36.712754 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzlv8\" (UniqueName: \"kubernetes.io/projected/806c9298-1aad-4e84-a80a-b2078f548d03-kube-api-access-lzlv8\") on node \"crc\" DevicePath \"\"" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.157128 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" event={"ID":"806c9298-1aad-4e84-a80a-b2078f548d03","Type":"ContainerDied","Data":"5ef0e14dae48201f24c7a94c3957558ad2320fc04d32e643522a9ea85874e0de"} Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.157172 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ef0e14dae48201f24c7a94c3957558ad2320fc04d32e643522a9ea85874e0de" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.157236 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-595lb" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.297116 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q"] Nov 22 05:41:37 crc kubenswrapper[4818]: E1122 05:41:37.297557 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b7bb1f-3d8c-41bb-8933-a9904a04b635" containerName="extract-utilities" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.297581 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b7bb1f-3d8c-41bb-8933-a9904a04b635" containerName="extract-utilities" Nov 22 05:41:37 crc kubenswrapper[4818]: E1122 05:41:37.297600 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="806c9298-1aad-4e84-a80a-b2078f548d03" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.297609 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="806c9298-1aad-4e84-a80a-b2078f548d03" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 22 05:41:37 crc kubenswrapper[4818]: E1122 05:41:37.297632 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b7bb1f-3d8c-41bb-8933-a9904a04b635" containerName="registry-server" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.297640 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b7bb1f-3d8c-41bb-8933-a9904a04b635" containerName="registry-server" Nov 22 05:41:37 crc kubenswrapper[4818]: E1122 05:41:37.297673 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b7bb1f-3d8c-41bb-8933-a9904a04b635" containerName="extract-content" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.297680 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b7bb1f-3d8c-41bb-8933-a9904a04b635" containerName="extract-content" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.297886 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="59b7bb1f-3d8c-41bb-8933-a9904a04b635" containerName="registry-server" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.297910 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="806c9298-1aad-4e84-a80a-b2078f548d03" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.298625 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.302370 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.302426 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.302635 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.303153 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fwnsc" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.303646 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.304233 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.304587 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.304913 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.309536 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.319609 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q"] Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.327436 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.327650 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.327831 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.327988 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm4kx\" (UniqueName: \"kubernetes.io/projected/38964063-2a2d-4728-9ccc-d5eca9db132d-kube-api-access-wm4kx\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.328196 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.328375 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.328572 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.328810 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.329864 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/38964063-2a2d-4728-9ccc-d5eca9db132d-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.329895 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.329945 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.431177 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.431280 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/38964063-2a2d-4728-9ccc-d5eca9db132d-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.431305 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.431350 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.431385 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.431424 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.431467 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.431504 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm4kx\" (UniqueName: \"kubernetes.io/projected/38964063-2a2d-4728-9ccc-d5eca9db132d-kube-api-access-wm4kx\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.431526 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.431548 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.431581 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.432389 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.432459 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/38964063-2a2d-4728-9ccc-d5eca9db132d-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.436700 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.436770 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.436966 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.437448 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.438007 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.438046 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.438585 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.441320 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.461182 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm4kx\" (UniqueName: \"kubernetes.io/projected/38964063-2a2d-4728-9ccc-d5eca9db132d-kube-api-access-wm4kx\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:37 crc kubenswrapper[4818]: I1122 05:41:37.630148 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:41:38 crc kubenswrapper[4818]: I1122 05:41:38.307388 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q"] Nov 22 05:41:39 crc kubenswrapper[4818]: I1122 05:41:39.182123 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" event={"ID":"38964063-2a2d-4728-9ccc-d5eca9db132d","Type":"ContainerStarted","Data":"8662e3368476c62499e54db688c12d1131d356c87d05c9bbe66f4af5106f03bd"} Nov 22 05:41:39 crc kubenswrapper[4818]: I1122 05:41:39.182539 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" event={"ID":"38964063-2a2d-4728-9ccc-d5eca9db132d","Type":"ContainerStarted","Data":"282a1665ec841157775f7fd9516f0c8e4b8017a3bada9b57630c2c406d54e573"} Nov 22 05:41:39 crc kubenswrapper[4818]: I1122 05:41:39.204481 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" podStartSLOduration=1.697596065 podStartE2EDuration="2.204454766s" podCreationTimestamp="2025-11-22 05:41:37 +0000 UTC" firstStartedPulling="2025-11-22 05:41:38.295495464 +0000 UTC m=+3250.869912021" lastFinishedPulling="2025-11-22 05:41:38.802354185 +0000 UTC m=+3251.376770722" observedRunningTime="2025-11-22 05:41:39.202509035 +0000 UTC m=+3251.776925602" watchObservedRunningTime="2025-11-22 05:41:39.204454766 +0000 UTC m=+3251.778871303" Nov 22 05:41:48 crc kubenswrapper[4818]: I1122 05:41:48.297052 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:41:48 crc kubenswrapper[4818]: E1122 05:41:48.297986 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:42:00 crc kubenswrapper[4818]: I1122 05:42:00.292844 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:42:00 crc kubenswrapper[4818]: E1122 05:42:00.294856 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:42:11 crc kubenswrapper[4818]: I1122 05:42:11.292717 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:42:11 crc kubenswrapper[4818]: E1122 05:42:11.293940 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:42:23 crc kubenswrapper[4818]: I1122 05:42:23.291802 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:42:23 crc kubenswrapper[4818]: E1122 05:42:23.293905 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:42:37 crc kubenswrapper[4818]: I1122 05:42:37.292508 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:42:37 crc kubenswrapper[4818]: E1122 05:42:37.293635 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:42:50 crc kubenswrapper[4818]: I1122 05:42:50.291636 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:42:50 crc kubenswrapper[4818]: E1122 05:42:50.292721 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:43:02 crc kubenswrapper[4818]: I1122 05:43:02.291865 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:43:02 crc kubenswrapper[4818]: E1122 05:43:02.292855 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:43:17 crc kubenswrapper[4818]: I1122 05:43:17.292202 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:43:17 crc kubenswrapper[4818]: E1122 05:43:17.293466 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:43:31 crc kubenswrapper[4818]: I1122 05:43:31.291405 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:43:32 crc kubenswrapper[4818]: I1122 05:43:32.464957 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"30eba2944bc07de93901d264862a9a44b1afc7503d2ddd03ddaec3e1e784bc41"} Nov 22 05:43:36 crc kubenswrapper[4818]: I1122 05:43:36.660171 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8669d"] Nov 22 05:43:36 crc kubenswrapper[4818]: I1122 05:43:36.662616 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8669d" Nov 22 05:43:36 crc kubenswrapper[4818]: I1122 05:43:36.669186 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdtqc\" (UniqueName: \"kubernetes.io/projected/48054291-f084-4c4e-803e-eeac1f65bed5-kube-api-access-kdtqc\") pod \"redhat-operators-8669d\" (UID: \"48054291-f084-4c4e-803e-eeac1f65bed5\") " pod="openshift-marketplace/redhat-operators-8669d" Nov 22 05:43:36 crc kubenswrapper[4818]: I1122 05:43:36.669668 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48054291-f084-4c4e-803e-eeac1f65bed5-catalog-content\") pod \"redhat-operators-8669d\" (UID: \"48054291-f084-4c4e-803e-eeac1f65bed5\") " pod="openshift-marketplace/redhat-operators-8669d" Nov 22 05:43:36 crc kubenswrapper[4818]: I1122 05:43:36.669728 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48054291-f084-4c4e-803e-eeac1f65bed5-utilities\") pod \"redhat-operators-8669d\" (UID: \"48054291-f084-4c4e-803e-eeac1f65bed5\") " pod="openshift-marketplace/redhat-operators-8669d" Nov 22 05:43:36 crc kubenswrapper[4818]: I1122 05:43:36.701093 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8669d"] Nov 22 05:43:36 crc kubenswrapper[4818]: I1122 05:43:36.771587 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48054291-f084-4c4e-803e-eeac1f65bed5-catalog-content\") pod \"redhat-operators-8669d\" (UID: \"48054291-f084-4c4e-803e-eeac1f65bed5\") " pod="openshift-marketplace/redhat-operators-8669d" Nov 22 05:43:36 crc kubenswrapper[4818]: I1122 05:43:36.771643 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48054291-f084-4c4e-803e-eeac1f65bed5-utilities\") pod \"redhat-operators-8669d\" (UID: \"48054291-f084-4c4e-803e-eeac1f65bed5\") " pod="openshift-marketplace/redhat-operators-8669d" Nov 22 05:43:36 crc kubenswrapper[4818]: I1122 05:43:36.771719 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdtqc\" (UniqueName: \"kubernetes.io/projected/48054291-f084-4c4e-803e-eeac1f65bed5-kube-api-access-kdtqc\") pod \"redhat-operators-8669d\" (UID: \"48054291-f084-4c4e-803e-eeac1f65bed5\") " pod="openshift-marketplace/redhat-operators-8669d" Nov 22 05:43:36 crc kubenswrapper[4818]: I1122 05:43:36.772176 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48054291-f084-4c4e-803e-eeac1f65bed5-catalog-content\") pod \"redhat-operators-8669d\" (UID: \"48054291-f084-4c4e-803e-eeac1f65bed5\") " pod="openshift-marketplace/redhat-operators-8669d" Nov 22 05:43:36 crc kubenswrapper[4818]: I1122 05:43:36.772480 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48054291-f084-4c4e-803e-eeac1f65bed5-utilities\") pod \"redhat-operators-8669d\" (UID: \"48054291-f084-4c4e-803e-eeac1f65bed5\") " pod="openshift-marketplace/redhat-operators-8669d" Nov 22 05:43:36 crc kubenswrapper[4818]: I1122 05:43:36.801418 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdtqc\" (UniqueName: \"kubernetes.io/projected/48054291-f084-4c4e-803e-eeac1f65bed5-kube-api-access-kdtqc\") pod \"redhat-operators-8669d\" (UID: \"48054291-f084-4c4e-803e-eeac1f65bed5\") " pod="openshift-marketplace/redhat-operators-8669d" Nov 22 05:43:36 crc kubenswrapper[4818]: I1122 05:43:36.991521 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8669d" Nov 22 05:43:37 crc kubenswrapper[4818]: I1122 05:43:37.489200 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8669d"] Nov 22 05:43:37 crc kubenswrapper[4818]: I1122 05:43:37.517697 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8669d" event={"ID":"48054291-f084-4c4e-803e-eeac1f65bed5","Type":"ContainerStarted","Data":"b69864a490164ffaadc4d3bff6b703d5388403cff9a57ea2896cdaed3f9139ac"} Nov 22 05:43:38 crc kubenswrapper[4818]: I1122 05:43:38.529642 4818 generic.go:334] "Generic (PLEG): container finished" podID="48054291-f084-4c4e-803e-eeac1f65bed5" containerID="38f75244489e609f45744bf4085f2bcda7b5a55431927d7cc7ab35465c075702" exitCode=0 Nov 22 05:43:38 crc kubenswrapper[4818]: I1122 05:43:38.529793 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8669d" event={"ID":"48054291-f084-4c4e-803e-eeac1f65bed5","Type":"ContainerDied","Data":"38f75244489e609f45744bf4085f2bcda7b5a55431927d7cc7ab35465c075702"} Nov 22 05:43:38 crc kubenswrapper[4818]: I1122 05:43:38.532431 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 05:43:40 crc kubenswrapper[4818]: I1122 05:43:40.554394 4818 generic.go:334] "Generic (PLEG): container finished" podID="48054291-f084-4c4e-803e-eeac1f65bed5" containerID="632919fcd847212ff6a73b36206348a904e856680a564f7277821b17d9887fa1" exitCode=0 Nov 22 05:43:40 crc kubenswrapper[4818]: I1122 05:43:40.554515 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8669d" event={"ID":"48054291-f084-4c4e-803e-eeac1f65bed5","Type":"ContainerDied","Data":"632919fcd847212ff6a73b36206348a904e856680a564f7277821b17d9887fa1"} Nov 22 05:43:41 crc kubenswrapper[4818]: I1122 05:43:41.572869 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8669d" event={"ID":"48054291-f084-4c4e-803e-eeac1f65bed5","Type":"ContainerStarted","Data":"e20d9840ff59dd7d8458d2b9710c2722a57e3d908d34bef0930e1d4277ba634d"} Nov 22 05:43:41 crc kubenswrapper[4818]: I1122 05:43:41.595671 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8669d" podStartSLOduration=3.182026713 podStartE2EDuration="5.595646753s" podCreationTimestamp="2025-11-22 05:43:36 +0000 UTC" firstStartedPulling="2025-11-22 05:43:38.532070336 +0000 UTC m=+3371.106486873" lastFinishedPulling="2025-11-22 05:43:40.945690376 +0000 UTC m=+3373.520106913" observedRunningTime="2025-11-22 05:43:41.593382932 +0000 UTC m=+3374.167799469" watchObservedRunningTime="2025-11-22 05:43:41.595646753 +0000 UTC m=+3374.170063290" Nov 22 05:43:46 crc kubenswrapper[4818]: I1122 05:43:46.992557 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8669d" Nov 22 05:43:46 crc kubenswrapper[4818]: I1122 05:43:46.993168 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8669d" Nov 22 05:43:48 crc kubenswrapper[4818]: I1122 05:43:48.056902 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8669d" podUID="48054291-f084-4c4e-803e-eeac1f65bed5" containerName="registry-server" probeResult="failure" output=< Nov 22 05:43:48 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Nov 22 05:43:48 crc kubenswrapper[4818]: > Nov 22 05:43:57 crc kubenswrapper[4818]: I1122 05:43:57.065521 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8669d" Nov 22 05:43:57 crc kubenswrapper[4818]: I1122 05:43:57.126059 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8669d" Nov 22 05:43:57 crc kubenswrapper[4818]: I1122 05:43:57.306872 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8669d"] Nov 22 05:43:58 crc kubenswrapper[4818]: I1122 05:43:58.758113 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8669d" podUID="48054291-f084-4c4e-803e-eeac1f65bed5" containerName="registry-server" containerID="cri-o://e20d9840ff59dd7d8458d2b9710c2722a57e3d908d34bef0930e1d4277ba634d" gracePeriod=2 Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.246707 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8669d" Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.348377 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdtqc\" (UniqueName: \"kubernetes.io/projected/48054291-f084-4c4e-803e-eeac1f65bed5-kube-api-access-kdtqc\") pod \"48054291-f084-4c4e-803e-eeac1f65bed5\" (UID: \"48054291-f084-4c4e-803e-eeac1f65bed5\") " Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.348786 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48054291-f084-4c4e-803e-eeac1f65bed5-utilities\") pod \"48054291-f084-4c4e-803e-eeac1f65bed5\" (UID: \"48054291-f084-4c4e-803e-eeac1f65bed5\") " Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.348831 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48054291-f084-4c4e-803e-eeac1f65bed5-catalog-content\") pod \"48054291-f084-4c4e-803e-eeac1f65bed5\" (UID: \"48054291-f084-4c4e-803e-eeac1f65bed5\") " Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.350644 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48054291-f084-4c4e-803e-eeac1f65bed5-utilities" (OuterVolumeSpecName: "utilities") pod "48054291-f084-4c4e-803e-eeac1f65bed5" (UID: "48054291-f084-4c4e-803e-eeac1f65bed5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.356459 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48054291-f084-4c4e-803e-eeac1f65bed5-kube-api-access-kdtqc" (OuterVolumeSpecName: "kube-api-access-kdtqc") pod "48054291-f084-4c4e-803e-eeac1f65bed5" (UID: "48054291-f084-4c4e-803e-eeac1f65bed5"). InnerVolumeSpecName "kube-api-access-kdtqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.451976 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdtqc\" (UniqueName: \"kubernetes.io/projected/48054291-f084-4c4e-803e-eeac1f65bed5-kube-api-access-kdtqc\") on node \"crc\" DevicePath \"\"" Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.452030 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48054291-f084-4c4e-803e-eeac1f65bed5-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.478683 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48054291-f084-4c4e-803e-eeac1f65bed5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48054291-f084-4c4e-803e-eeac1f65bed5" (UID: "48054291-f084-4c4e-803e-eeac1f65bed5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.553878 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48054291-f084-4c4e-803e-eeac1f65bed5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.767152 4818 generic.go:334] "Generic (PLEG): container finished" podID="48054291-f084-4c4e-803e-eeac1f65bed5" containerID="e20d9840ff59dd7d8458d2b9710c2722a57e3d908d34bef0930e1d4277ba634d" exitCode=0 Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.767202 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8669d" Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.767217 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8669d" event={"ID":"48054291-f084-4c4e-803e-eeac1f65bed5","Type":"ContainerDied","Data":"e20d9840ff59dd7d8458d2b9710c2722a57e3d908d34bef0930e1d4277ba634d"} Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.767276 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8669d" event={"ID":"48054291-f084-4c4e-803e-eeac1f65bed5","Type":"ContainerDied","Data":"b69864a490164ffaadc4d3bff6b703d5388403cff9a57ea2896cdaed3f9139ac"} Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.767297 4818 scope.go:117] "RemoveContainer" containerID="e20d9840ff59dd7d8458d2b9710c2722a57e3d908d34bef0930e1d4277ba634d" Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.796673 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8669d"] Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.800718 4818 scope.go:117] "RemoveContainer" containerID="632919fcd847212ff6a73b36206348a904e856680a564f7277821b17d9887fa1" Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.805653 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8669d"] Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.819222 4818 scope.go:117] "RemoveContainer" containerID="38f75244489e609f45744bf4085f2bcda7b5a55431927d7cc7ab35465c075702" Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.857850 4818 scope.go:117] "RemoveContainer" containerID="e20d9840ff59dd7d8458d2b9710c2722a57e3d908d34bef0930e1d4277ba634d" Nov 22 05:43:59 crc kubenswrapper[4818]: E1122 05:43:59.858348 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e20d9840ff59dd7d8458d2b9710c2722a57e3d908d34bef0930e1d4277ba634d\": container with ID starting with e20d9840ff59dd7d8458d2b9710c2722a57e3d908d34bef0930e1d4277ba634d not found: ID does not exist" containerID="e20d9840ff59dd7d8458d2b9710c2722a57e3d908d34bef0930e1d4277ba634d" Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.858402 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e20d9840ff59dd7d8458d2b9710c2722a57e3d908d34bef0930e1d4277ba634d"} err="failed to get container status \"e20d9840ff59dd7d8458d2b9710c2722a57e3d908d34bef0930e1d4277ba634d\": rpc error: code = NotFound desc = could not find container \"e20d9840ff59dd7d8458d2b9710c2722a57e3d908d34bef0930e1d4277ba634d\": container with ID starting with e20d9840ff59dd7d8458d2b9710c2722a57e3d908d34bef0930e1d4277ba634d not found: ID does not exist" Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.858444 4818 scope.go:117] "RemoveContainer" containerID="632919fcd847212ff6a73b36206348a904e856680a564f7277821b17d9887fa1" Nov 22 05:43:59 crc kubenswrapper[4818]: E1122 05:43:59.858902 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"632919fcd847212ff6a73b36206348a904e856680a564f7277821b17d9887fa1\": container with ID starting with 632919fcd847212ff6a73b36206348a904e856680a564f7277821b17d9887fa1 not found: ID does not exist" containerID="632919fcd847212ff6a73b36206348a904e856680a564f7277821b17d9887fa1" Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.858932 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"632919fcd847212ff6a73b36206348a904e856680a564f7277821b17d9887fa1"} err="failed to get container status \"632919fcd847212ff6a73b36206348a904e856680a564f7277821b17d9887fa1\": rpc error: code = NotFound desc = could not find container \"632919fcd847212ff6a73b36206348a904e856680a564f7277821b17d9887fa1\": container with ID starting with 632919fcd847212ff6a73b36206348a904e856680a564f7277821b17d9887fa1 not found: ID does not exist" Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.858955 4818 scope.go:117] "RemoveContainer" containerID="38f75244489e609f45744bf4085f2bcda7b5a55431927d7cc7ab35465c075702" Nov 22 05:43:59 crc kubenswrapper[4818]: E1122 05:43:59.859197 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38f75244489e609f45744bf4085f2bcda7b5a55431927d7cc7ab35465c075702\": container with ID starting with 38f75244489e609f45744bf4085f2bcda7b5a55431927d7cc7ab35465c075702 not found: ID does not exist" containerID="38f75244489e609f45744bf4085f2bcda7b5a55431927d7cc7ab35465c075702" Nov 22 05:43:59 crc kubenswrapper[4818]: I1122 05:43:59.859216 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38f75244489e609f45744bf4085f2bcda7b5a55431927d7cc7ab35465c075702"} err="failed to get container status \"38f75244489e609f45744bf4085f2bcda7b5a55431927d7cc7ab35465c075702\": rpc error: code = NotFound desc = could not find container \"38f75244489e609f45744bf4085f2bcda7b5a55431927d7cc7ab35465c075702\": container with ID starting with 38f75244489e609f45744bf4085f2bcda7b5a55431927d7cc7ab35465c075702 not found: ID does not exist" Nov 22 05:44:00 crc kubenswrapper[4818]: I1122 05:44:00.303566 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48054291-f084-4c4e-803e-eeac1f65bed5" path="/var/lib/kubelet/pods/48054291-f084-4c4e-803e-eeac1f65bed5/volumes" Nov 22 05:45:00 crc kubenswrapper[4818]: I1122 05:45:00.170755 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396505-p7r4h"] Nov 22 05:45:00 crc kubenswrapper[4818]: E1122 05:45:00.171778 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48054291-f084-4c4e-803e-eeac1f65bed5" containerName="registry-server" Nov 22 05:45:00 crc kubenswrapper[4818]: I1122 05:45:00.171794 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="48054291-f084-4c4e-803e-eeac1f65bed5" containerName="registry-server" Nov 22 05:45:00 crc kubenswrapper[4818]: E1122 05:45:00.171829 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48054291-f084-4c4e-803e-eeac1f65bed5" containerName="extract-content" Nov 22 05:45:00 crc kubenswrapper[4818]: I1122 05:45:00.171838 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="48054291-f084-4c4e-803e-eeac1f65bed5" containerName="extract-content" Nov 22 05:45:00 crc kubenswrapper[4818]: E1122 05:45:00.171869 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48054291-f084-4c4e-803e-eeac1f65bed5" containerName="extract-utilities" Nov 22 05:45:00 crc kubenswrapper[4818]: I1122 05:45:00.171877 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="48054291-f084-4c4e-803e-eeac1f65bed5" containerName="extract-utilities" Nov 22 05:45:00 crc kubenswrapper[4818]: I1122 05:45:00.172074 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="48054291-f084-4c4e-803e-eeac1f65bed5" containerName="registry-server" Nov 22 05:45:00 crc kubenswrapper[4818]: I1122 05:45:00.172827 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396505-p7r4h" Nov 22 05:45:00 crc kubenswrapper[4818]: I1122 05:45:00.178589 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 05:45:00 crc kubenswrapper[4818]: I1122 05:45:00.179953 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 05:45:00 crc kubenswrapper[4818]: I1122 05:45:00.183312 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396505-p7r4h"] Nov 22 05:45:00 crc kubenswrapper[4818]: I1122 05:45:00.316629 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e-config-volume\") pod \"collect-profiles-29396505-p7r4h\" (UID: \"40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396505-p7r4h" Nov 22 05:45:00 crc kubenswrapper[4818]: I1122 05:45:00.316898 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e-secret-volume\") pod \"collect-profiles-29396505-p7r4h\" (UID: \"40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396505-p7r4h" Nov 22 05:45:00 crc kubenswrapper[4818]: I1122 05:45:00.317235 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m62xh\" (UniqueName: \"kubernetes.io/projected/40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e-kube-api-access-m62xh\") pod \"collect-profiles-29396505-p7r4h\" (UID: \"40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396505-p7r4h" Nov 22 05:45:00 crc kubenswrapper[4818]: I1122 05:45:00.418569 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m62xh\" (UniqueName: \"kubernetes.io/projected/40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e-kube-api-access-m62xh\") pod \"collect-profiles-29396505-p7r4h\" (UID: \"40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396505-p7r4h" Nov 22 05:45:00 crc kubenswrapper[4818]: I1122 05:45:00.419075 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e-config-volume\") pod \"collect-profiles-29396505-p7r4h\" (UID: \"40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396505-p7r4h" Nov 22 05:45:00 crc kubenswrapper[4818]: I1122 05:45:00.420202 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e-config-volume\") pod \"collect-profiles-29396505-p7r4h\" (UID: \"40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396505-p7r4h" Nov 22 05:45:00 crc kubenswrapper[4818]: I1122 05:45:00.443772 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e-secret-volume\") pod \"collect-profiles-29396505-p7r4h\" (UID: \"40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396505-p7r4h" Nov 22 05:45:00 crc kubenswrapper[4818]: I1122 05:45:00.448815 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e-secret-volume\") pod \"collect-profiles-29396505-p7r4h\" (UID: \"40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396505-p7r4h" Nov 22 05:45:00 crc kubenswrapper[4818]: I1122 05:45:00.456105 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m62xh\" (UniqueName: \"kubernetes.io/projected/40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e-kube-api-access-m62xh\") pod \"collect-profiles-29396505-p7r4h\" (UID: \"40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396505-p7r4h" Nov 22 05:45:00 crc kubenswrapper[4818]: I1122 05:45:00.493693 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396505-p7r4h" Nov 22 05:45:00 crc kubenswrapper[4818]: I1122 05:45:00.953344 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396505-p7r4h"] Nov 22 05:45:01 crc kubenswrapper[4818]: I1122 05:45:01.456845 4818 generic.go:334] "Generic (PLEG): container finished" podID="40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e" containerID="2b16e0edcfc0615feca7e7c9d746d2a4a04e687ddeb1fecde93014145ceafaf8" exitCode=0 Nov 22 05:45:01 crc kubenswrapper[4818]: I1122 05:45:01.456963 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396505-p7r4h" event={"ID":"40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e","Type":"ContainerDied","Data":"2b16e0edcfc0615feca7e7c9d746d2a4a04e687ddeb1fecde93014145ceafaf8"} Nov 22 05:45:01 crc kubenswrapper[4818]: I1122 05:45:01.458115 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396505-p7r4h" event={"ID":"40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e","Type":"ContainerStarted","Data":"f4b5269b759917aea5b95d5dc616618f803b3862d4bb365342ad529f0405e73b"} Nov 22 05:45:02 crc kubenswrapper[4818]: I1122 05:45:02.850785 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396505-p7r4h" Nov 22 05:45:02 crc kubenswrapper[4818]: I1122 05:45:02.893128 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e-secret-volume\") pod \"40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e\" (UID: \"40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e\") " Nov 22 05:45:02 crc kubenswrapper[4818]: I1122 05:45:02.893206 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e-config-volume\") pod \"40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e\" (UID: \"40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e\") " Nov 22 05:45:02 crc kubenswrapper[4818]: I1122 05:45:02.893328 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m62xh\" (UniqueName: \"kubernetes.io/projected/40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e-kube-api-access-m62xh\") pod \"40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e\" (UID: \"40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e\") " Nov 22 05:45:02 crc kubenswrapper[4818]: I1122 05:45:02.895062 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e-config-volume" (OuterVolumeSpecName: "config-volume") pod "40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e" (UID: "40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:45:02 crc kubenswrapper[4818]: I1122 05:45:02.900639 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e" (UID: "40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:02 crc kubenswrapper[4818]: I1122 05:45:02.900733 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e-kube-api-access-m62xh" (OuterVolumeSpecName: "kube-api-access-m62xh") pod "40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e" (UID: "40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e"). InnerVolumeSpecName "kube-api-access-m62xh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:45:02 crc kubenswrapper[4818]: I1122 05:45:02.995275 4818 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:02 crc kubenswrapper[4818]: I1122 05:45:02.995318 4818 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:02 crc kubenswrapper[4818]: I1122 05:45:02.995335 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m62xh\" (UniqueName: \"kubernetes.io/projected/40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e-kube-api-access-m62xh\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:03 crc kubenswrapper[4818]: I1122 05:45:03.480290 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396505-p7r4h" event={"ID":"40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e","Type":"ContainerDied","Data":"f4b5269b759917aea5b95d5dc616618f803b3862d4bb365342ad529f0405e73b"} Nov 22 05:45:03 crc kubenswrapper[4818]: I1122 05:45:03.480616 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4b5269b759917aea5b95d5dc616618f803b3862d4bb365342ad529f0405e73b" Nov 22 05:45:03 crc kubenswrapper[4818]: I1122 05:45:03.480406 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396505-p7r4h" Nov 22 05:45:03 crc kubenswrapper[4818]: I1122 05:45:03.924502 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq"] Nov 22 05:45:03 crc kubenswrapper[4818]: I1122 05:45:03.930811 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396460-hb4dq"] Nov 22 05:45:04 crc kubenswrapper[4818]: I1122 05:45:04.306214 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="176f1072-37ec-476e-8bcc-eda2271dd884" path="/var/lib/kubelet/pods/176f1072-37ec-476e-8bcc-eda2271dd884/volumes" Nov 22 05:45:11 crc kubenswrapper[4818]: E1122 05:45:11.053165 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38964063_2a2d_4728_9ccc_d5eca9db132d.slice/crio-conmon-8662e3368476c62499e54db688c12d1131d356c87d05c9bbe66f4af5106f03bd.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38964063_2a2d_4728_9ccc_d5eca9db132d.slice/crio-8662e3368476c62499e54db688c12d1131d356c87d05c9bbe66f4af5106f03bd.scope\": RecentStats: unable to find data in memory cache]" Nov 22 05:45:11 crc kubenswrapper[4818]: I1122 05:45:11.562802 4818 generic.go:334] "Generic (PLEG): container finished" podID="38964063-2a2d-4728-9ccc-d5eca9db132d" containerID="8662e3368476c62499e54db688c12d1131d356c87d05c9bbe66f4af5106f03bd" exitCode=0 Nov 22 05:45:11 crc kubenswrapper[4818]: I1122 05:45:11.562865 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" event={"ID":"38964063-2a2d-4728-9ccc-d5eca9db132d","Type":"ContainerDied","Data":"8662e3368476c62499e54db688c12d1131d356c87d05c9bbe66f4af5106f03bd"} Nov 22 05:45:12 crc kubenswrapper[4818]: I1122 05:45:12.988628 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.101264 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-extra-config-0\") pod \"38964063-2a2d-4728-9ccc-d5eca9db132d\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.101353 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/38964063-2a2d-4728-9ccc-d5eca9db132d-ceph-nova-0\") pod \"38964063-2a2d-4728-9ccc-d5eca9db132d\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.101477 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-migration-ssh-key-0\") pod \"38964063-2a2d-4728-9ccc-d5eca9db132d\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.101497 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-migration-ssh-key-1\") pod \"38964063-2a2d-4728-9ccc-d5eca9db132d\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.101541 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-ssh-key\") pod \"38964063-2a2d-4728-9ccc-d5eca9db132d\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.101574 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-inventory\") pod \"38964063-2a2d-4728-9ccc-d5eca9db132d\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.101626 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-custom-ceph-combined-ca-bundle\") pod \"38964063-2a2d-4728-9ccc-d5eca9db132d\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.101664 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wm4kx\" (UniqueName: \"kubernetes.io/projected/38964063-2a2d-4728-9ccc-d5eca9db132d-kube-api-access-wm4kx\") pod \"38964063-2a2d-4728-9ccc-d5eca9db132d\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.101693 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-cell1-compute-config-1\") pod \"38964063-2a2d-4728-9ccc-d5eca9db132d\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.101786 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-cell1-compute-config-0\") pod \"38964063-2a2d-4728-9ccc-d5eca9db132d\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.101811 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-ceph\") pod \"38964063-2a2d-4728-9ccc-d5eca9db132d\" (UID: \"38964063-2a2d-4728-9ccc-d5eca9db132d\") " Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.107490 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "38964063-2a2d-4728-9ccc-d5eca9db132d" (UID: "38964063-2a2d-4728-9ccc-d5eca9db132d"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.108423 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38964063-2a2d-4728-9ccc-d5eca9db132d-kube-api-access-wm4kx" (OuterVolumeSpecName: "kube-api-access-wm4kx") pod "38964063-2a2d-4728-9ccc-d5eca9db132d" (UID: "38964063-2a2d-4728-9ccc-d5eca9db132d"). InnerVolumeSpecName "kube-api-access-wm4kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.109063 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-ceph" (OuterVolumeSpecName: "ceph") pod "38964063-2a2d-4728-9ccc-d5eca9db132d" (UID: "38964063-2a2d-4728-9ccc-d5eca9db132d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.126679 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "38964063-2a2d-4728-9ccc-d5eca9db132d" (UID: "38964063-2a2d-4728-9ccc-d5eca9db132d"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.130224 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "38964063-2a2d-4728-9ccc-d5eca9db132d" (UID: "38964063-2a2d-4728-9ccc-d5eca9db132d"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.134408 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "38964063-2a2d-4728-9ccc-d5eca9db132d" (UID: "38964063-2a2d-4728-9ccc-d5eca9db132d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.139912 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "38964063-2a2d-4728-9ccc-d5eca9db132d" (UID: "38964063-2a2d-4728-9ccc-d5eca9db132d"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.142590 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-inventory" (OuterVolumeSpecName: "inventory") pod "38964063-2a2d-4728-9ccc-d5eca9db132d" (UID: "38964063-2a2d-4728-9ccc-d5eca9db132d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.149539 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "38964063-2a2d-4728-9ccc-d5eca9db132d" (UID: "38964063-2a2d-4728-9ccc-d5eca9db132d"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.160211 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38964063-2a2d-4728-9ccc-d5eca9db132d-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "38964063-2a2d-4728-9ccc-d5eca9db132d" (UID: "38964063-2a2d-4728-9ccc-d5eca9db132d"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.173855 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "38964063-2a2d-4728-9ccc-d5eca9db132d" (UID: "38964063-2a2d-4728-9ccc-d5eca9db132d"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.205492 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.205552 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.205613 4818 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.205628 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wm4kx\" (UniqueName: \"kubernetes.io/projected/38964063-2a2d-4728-9ccc-d5eca9db132d-kube-api-access-wm4kx\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.205645 4818 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.205656 4818 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.205693 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.205908 4818 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.205947 4818 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/38964063-2a2d-4728-9ccc-d5eca9db132d-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.205958 4818 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.205969 4818 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/38964063-2a2d-4728-9ccc-d5eca9db132d-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.584372 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" event={"ID":"38964063-2a2d-4728-9ccc-d5eca9db132d","Type":"ContainerDied","Data":"282a1665ec841157775f7fd9516f0c8e4b8017a3bada9b57630c2c406d54e573"} Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.584433 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="282a1665ec841157775f7fd9516f0c8e4b8017a3bada9b57630c2c406d54e573" Nov 22 05:45:13 crc kubenswrapper[4818]: I1122 05:45:13.584534 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.077120 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 22 05:45:27 crc kubenswrapper[4818]: E1122 05:45:27.078091 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38964063-2a2d-4728-9ccc-d5eca9db132d" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.078109 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="38964063-2a2d-4728-9ccc-d5eca9db132d" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 22 05:45:27 crc kubenswrapper[4818]: E1122 05:45:27.078134 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e" containerName="collect-profiles" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.078140 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e" containerName="collect-profiles" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.078314 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="38964063-2a2d-4728-9ccc-d5eca9db132d" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.078343 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="40f8fe72-1daa-4eef-8bb0-3ffdaee9d45e" containerName="collect-profiles" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.079907 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.082970 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.083192 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.090842 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.172305 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.175620 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.179396 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.198195 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.224001 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ksfkm"] Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.226948 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ksfkm" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.242925 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ksfkm"] Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.274472 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dfef980-4eab-42b9-821e-233a264843cf-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.274513 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.274531 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-run\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.274551 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zl4l\" (UniqueName: \"kubernetes.io/projected/3dfef980-4eab-42b9-821e-233a264843cf-kube-api-access-6zl4l\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.274578 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3dfef980-4eab-42b9-821e-233a264843cf-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.274603 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dfef980-4eab-42b9-821e-233a264843cf-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.274627 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.274652 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3dfef980-4eab-42b9-821e-233a264843cf-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.274671 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-sys\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.274688 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.274708 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dfef980-4eab-42b9-821e-233a264843cf-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.274724 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.274736 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-dev\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.274752 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.274766 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.274800 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376082 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwwfk\" (UniqueName: \"kubernetes.io/projected/0c1e7078-787f-4725-80b1-45f7047e108f-kube-api-access-nwwfk\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376141 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-run\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376161 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c1e7078-787f-4725-80b1-45f7047e108f-config-data\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376213 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376239 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dfef980-4eab-42b9-821e-233a264843cf-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376274 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376290 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-run\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376312 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zl4l\" (UniqueName: \"kubernetes.io/projected/3dfef980-4eab-42b9-821e-233a264843cf-kube-api-access-6zl4l\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376330 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376355 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3759d20-d679-42de-9a4e-f32305ec1ca0-utilities\") pod \"certified-operators-ksfkm\" (UID: \"e3759d20-d679-42de-9a4e-f32305ec1ca0\") " pod="openshift-marketplace/certified-operators-ksfkm" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376372 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3dfef980-4eab-42b9-821e-233a264843cf-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376393 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376412 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dfef980-4eab-42b9-821e-233a264843cf-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376442 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376461 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9xxb\" (UniqueName: \"kubernetes.io/projected/e3759d20-d679-42de-9a4e-f32305ec1ca0-kube-api-access-q9xxb\") pod \"certified-operators-ksfkm\" (UID: \"e3759d20-d679-42de-9a4e-f32305ec1ca0\") " pod="openshift-marketplace/certified-operators-ksfkm" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376477 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c1e7078-787f-4725-80b1-45f7047e108f-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376494 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c1e7078-787f-4725-80b1-45f7047e108f-scripts\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376515 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3dfef980-4eab-42b9-821e-233a264843cf-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376534 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-sys\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376550 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376569 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-dev\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376584 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376602 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dfef980-4eab-42b9-821e-233a264843cf-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376622 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-dev\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376643 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376670 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-sys\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376693 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-lib-modules\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376713 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3759d20-d679-42de-9a4e-f32305ec1ca0-catalog-content\") pod \"certified-operators-ksfkm\" (UID: \"e3759d20-d679-42de-9a4e-f32305ec1ca0\") " pod="openshift-marketplace/certified-operators-ksfkm" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376739 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376764 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376788 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376823 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0c1e7078-787f-4725-80b1-45f7047e108f-ceph\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376852 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c1e7078-787f-4725-80b1-45f7047e108f-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376879 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.376906 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.377337 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.377562 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-dev\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.377621 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-sys\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.377716 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.378239 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.378376 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.378338 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.378543 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.378561 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.378573 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3dfef980-4eab-42b9-821e-233a264843cf-run\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.386129 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3dfef980-4eab-42b9-821e-233a264843cf-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.386209 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3dfef980-4eab-42b9-821e-233a264843cf-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.386423 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dfef980-4eab-42b9-821e-233a264843cf-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.387185 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dfef980-4eab-42b9-821e-233a264843cf-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.388563 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dfef980-4eab-42b9-821e-233a264843cf-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.400866 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zl4l\" (UniqueName: \"kubernetes.io/projected/3dfef980-4eab-42b9-821e-233a264843cf-kube-api-access-6zl4l\") pod \"cinder-volume-volume1-0\" (UID: \"3dfef980-4eab-42b9-821e-233a264843cf\") " pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.477838 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.478127 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.478182 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3759d20-d679-42de-9a4e-f32305ec1ca0-utilities\") pod \"certified-operators-ksfkm\" (UID: \"e3759d20-d679-42de-9a4e-f32305ec1ca0\") " pod="openshift-marketplace/certified-operators-ksfkm" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.478226 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.478334 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c1e7078-787f-4725-80b1-45f7047e108f-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.478358 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9xxb\" (UniqueName: \"kubernetes.io/projected/e3759d20-d679-42de-9a4e-f32305ec1ca0-kube-api-access-q9xxb\") pod \"certified-operators-ksfkm\" (UID: \"e3759d20-d679-42de-9a4e-f32305ec1ca0\") " pod="openshift-marketplace/certified-operators-ksfkm" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.478382 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c1e7078-787f-4725-80b1-45f7047e108f-scripts\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.478442 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-dev\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.478473 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.478506 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-sys\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.478526 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-lib-modules\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.478552 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3759d20-d679-42de-9a4e-f32305ec1ca0-catalog-content\") pod \"certified-operators-ksfkm\" (UID: \"e3759d20-d679-42de-9a4e-f32305ec1ca0\") " pod="openshift-marketplace/certified-operators-ksfkm" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.478577 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.478628 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0c1e7078-787f-4725-80b1-45f7047e108f-ceph\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.478657 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c1e7078-787f-4725-80b1-45f7047e108f-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.478683 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.478715 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwwfk\" (UniqueName: \"kubernetes.io/projected/0c1e7078-787f-4725-80b1-45f7047e108f-kube-api-access-nwwfk\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.478748 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-run\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.478770 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c1e7078-787f-4725-80b1-45f7047e108f-config-data\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.479141 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.477992 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.479396 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.479472 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-dev\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.479497 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3759d20-d679-42de-9a4e-f32305ec1ca0-utilities\") pod \"certified-operators-ksfkm\" (UID: \"e3759d20-d679-42de-9a4e-f32305ec1ca0\") " pod="openshift-marketplace/certified-operators-ksfkm" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.479364 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.479547 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.479995 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-sys\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.480060 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-run\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.480060 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.480075 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3759d20-d679-42de-9a4e-f32305ec1ca0-catalog-content\") pod \"certified-operators-ksfkm\" (UID: \"e3759d20-d679-42de-9a4e-f32305ec1ca0\") " pod="openshift-marketplace/certified-operators-ksfkm" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.480395 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0c1e7078-787f-4725-80b1-45f7047e108f-lib-modules\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.482936 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c1e7078-787f-4725-80b1-45f7047e108f-scripts\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.483092 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c1e7078-787f-4725-80b1-45f7047e108f-config-data\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.483970 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0c1e7078-787f-4725-80b1-45f7047e108f-ceph\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.492504 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c1e7078-787f-4725-80b1-45f7047e108f-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.492572 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c1e7078-787f-4725-80b1-45f7047e108f-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.509654 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9xxb\" (UniqueName: \"kubernetes.io/projected/e3759d20-d679-42de-9a4e-f32305ec1ca0-kube-api-access-q9xxb\") pod \"certified-operators-ksfkm\" (UID: \"e3759d20-d679-42de-9a4e-f32305ec1ca0\") " pod="openshift-marketplace/certified-operators-ksfkm" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.514103 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwwfk\" (UniqueName: \"kubernetes.io/projected/0c1e7078-787f-4725-80b1-45f7047e108f-kube-api-access-nwwfk\") pod \"cinder-backup-0\" (UID: \"0c1e7078-787f-4725-80b1-45f7047e108f\") " pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.516691 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.546065 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ksfkm" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.696920 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.911374 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-50ec-account-create-zq678"] Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.913118 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-50ec-account-create-zq678" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.923466 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.936874 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-50ec-account-create-zq678"] Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.975695 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-v2bv5"] Nov 22 05:45:27 crc kubenswrapper[4818]: I1122 05:45:27.979637 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-v2bv5" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.014806 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-v2bv5"] Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.049036 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.050863 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.060486 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-f784b6cfc-t55xf"] Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.061882 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.062053 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2zbds" Nov 22 05:45:28 crc kubenswrapper[4818]: W1122 05:45:28.062249 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3759d20_d679_42de_9a4e_f32305ec1ca0.slice/crio-111f708a9cacde7583bf2e26ada8a9d1460f6d13ff5ffd9b987c416bf415431a WatchSource:0}: Error finding container 111f708a9cacde7583bf2e26ada8a9d1460f6d13ff5ffd9b987c416bf415431a: Status 404 returned error can't find the container with id 111f708a9cacde7583bf2e26ada8a9d1460f6d13ff5ffd9b987c416bf415431a Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.062304 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.062355 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.062314 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f784b6cfc-t55xf" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.072978 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.073172 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.073393 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.073652 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-psw4h" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.093394 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bds4t\" (UniqueName: \"kubernetes.io/projected/1ca6675b-db38-49ba-a7ab-396d5d6dc54f-kube-api-access-bds4t\") pod \"manila-50ec-account-create-zq678\" (UID: \"1ca6675b-db38-49ba-a7ab-396d5d6dc54f\") " pod="openstack/manila-50ec-account-create-zq678" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.093461 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ca6675b-db38-49ba-a7ab-396d5d6dc54f-operator-scripts\") pod \"manila-50ec-account-create-zq678\" (UID: \"1ca6675b-db38-49ba-a7ab-396d5d6dc54f\") " pod="openstack/manila-50ec-account-create-zq678" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.093567 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/384fe684-9694-4069-b37e-ad50f83ef867-operator-scripts\") pod \"manila-db-create-v2bv5\" (UID: \"384fe684-9694-4069-b37e-ad50f83ef867\") " pod="openstack/manila-db-create-v2bv5" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.093583 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfh7r\" (UniqueName: \"kubernetes.io/projected/384fe684-9694-4069-b37e-ad50f83ef867-kube-api-access-dfh7r\") pod \"manila-db-create-v2bv5\" (UID: \"384fe684-9694-4069-b37e-ad50f83ef867\") " pod="openstack/manila-db-create-v2bv5" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.100071 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ksfkm"] Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.118617 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 05:45:28 crc kubenswrapper[4818]: E1122 05:45:28.120516 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceph combined-ca-bundle config-data glance httpd-run kube-api-access-hf46n logs public-tls-certs scripts], unattached volumes=[], failed to process volumes=[ceph combined-ca-bundle config-data glance httpd-run kube-api-access-hf46n logs public-tls-certs scripts]: context canceled" pod="openstack/glance-default-external-api-0" podUID="244410df-3345-4aa9-8d98-677e5593bfad" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.168666 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f784b6cfc-t55xf"] Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.192224 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.193752 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.196005 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/244410df-3345-4aa9-8d98-677e5593bfad-ceph\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.196106 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ca6675b-db38-49ba-a7ab-396d5d6dc54f-operator-scripts\") pod \"manila-50ec-account-create-zq678\" (UID: \"1ca6675b-db38-49ba-a7ab-396d5d6dc54f\") " pod="openstack/manila-50ec-account-create-zq678" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.196140 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d03bb0b9-32cd-43b8-92e4-2b8f83361246-config-data\") pod \"horizon-f784b6cfc-t55xf\" (UID: \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\") " pod="openstack/horizon-f784b6cfc-t55xf" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.196157 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d03bb0b9-32cd-43b8-92e4-2b8f83361246-horizon-secret-key\") pod \"horizon-f784b6cfc-t55xf\" (UID: \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\") " pod="openstack/horizon-f784b6cfc-t55xf" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.196176 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf46n\" (UniqueName: \"kubernetes.io/projected/244410df-3345-4aa9-8d98-677e5593bfad-kube-api-access-hf46n\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.196378 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.196413 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d03bb0b9-32cd-43b8-92e4-2b8f83361246-logs\") pod \"horizon-f784b6cfc-t55xf\" (UID: \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\") " pod="openstack/horizon-f784b6cfc-t55xf" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.196432 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/244410df-3345-4aa9-8d98-677e5593bfad-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.196457 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.196471 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/244410df-3345-4aa9-8d98-677e5593bfad-logs\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.196525 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d03bb0b9-32cd-43b8-92e4-2b8f83361246-scripts\") pod \"horizon-f784b6cfc-t55xf\" (UID: \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\") " pod="openstack/horizon-f784b6cfc-t55xf" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.196549 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh594\" (UniqueName: \"kubernetes.io/projected/d03bb0b9-32cd-43b8-92e4-2b8f83361246-kube-api-access-dh594\") pod \"horizon-f784b6cfc-t55xf\" (UID: \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\") " pod="openstack/horizon-f784b6cfc-t55xf" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.196577 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/384fe684-9694-4069-b37e-ad50f83ef867-operator-scripts\") pod \"manila-db-create-v2bv5\" (UID: \"384fe684-9694-4069-b37e-ad50f83ef867\") " pod="openstack/manila-db-create-v2bv5" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.196592 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfh7r\" (UniqueName: \"kubernetes.io/projected/384fe684-9694-4069-b37e-ad50f83ef867-kube-api-access-dfh7r\") pod \"manila-db-create-v2bv5\" (UID: \"384fe684-9694-4069-b37e-ad50f83ef867\") " pod="openstack/manila-db-create-v2bv5" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.196611 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-config-data\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.196627 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.196653 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bds4t\" (UniqueName: \"kubernetes.io/projected/1ca6675b-db38-49ba-a7ab-396d5d6dc54f-kube-api-access-bds4t\") pod \"manila-50ec-account-create-zq678\" (UID: \"1ca6675b-db38-49ba-a7ab-396d5d6dc54f\") " pod="openstack/manila-50ec-account-create-zq678" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.196667 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-scripts\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.197412 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ca6675b-db38-49ba-a7ab-396d5d6dc54f-operator-scripts\") pod \"manila-50ec-account-create-zq678\" (UID: \"1ca6675b-db38-49ba-a7ab-396d5d6dc54f\") " pod="openstack/manila-50ec-account-create-zq678" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.198034 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/384fe684-9694-4069-b37e-ad50f83ef867-operator-scripts\") pod \"manila-db-create-v2bv5\" (UID: \"384fe684-9694-4069-b37e-ad50f83ef867\") " pod="openstack/manila-db-create-v2bv5" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.199541 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.199708 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.202989 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.227313 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfh7r\" (UniqueName: \"kubernetes.io/projected/384fe684-9694-4069-b37e-ad50f83ef867-kube-api-access-dfh7r\") pod \"manila-db-create-v2bv5\" (UID: \"384fe684-9694-4069-b37e-ad50f83ef867\") " pod="openstack/manila-db-create-v2bv5" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.227399 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bds4t\" (UniqueName: \"kubernetes.io/projected/1ca6675b-db38-49ba-a7ab-396d5d6dc54f-kube-api-access-bds4t\") pod \"manila-50ec-account-create-zq678\" (UID: \"1ca6675b-db38-49ba-a7ab-396d5d6dc54f\") " pod="openstack/manila-50ec-account-create-zq678" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.227673 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.239675 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.249659 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-50ec-account-create-zq678" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.260801 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-c4458788c-rwwjv"] Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.262383 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c4458788c-rwwjv" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.280427 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c4458788c-rwwjv"] Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.285872 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 05:45:28 crc kubenswrapper[4818]: E1122 05:45:28.286867 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceph combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-h5zgk logs scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-internal-api-0" podUID="26dce0f9-35d4-47e8-8bdc-61a508050489" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.319654 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-v2bv5" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.323362 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.327425 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d03bb0b9-32cd-43b8-92e4-2b8f83361246-config-data\") pod \"horizon-f784b6cfc-t55xf\" (UID: \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\") " pod="openstack/horizon-f784b6cfc-t55xf" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.327481 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d03bb0b9-32cd-43b8-92e4-2b8f83361246-horizon-secret-key\") pod \"horizon-f784b6cfc-t55xf\" (UID: \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\") " pod="openstack/horizon-f784b6cfc-t55xf" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.327511 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26dce0f9-35d4-47e8-8bdc-61a508050489-logs\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.327634 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf46n\" (UniqueName: \"kubernetes.io/projected/244410df-3345-4aa9-8d98-677e5593bfad-kube-api-access-hf46n\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.327705 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.327735 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26dce0f9-35d4-47e8-8bdc-61a508050489-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.327784 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d03bb0b9-32cd-43b8-92e4-2b8f83361246-logs\") pod \"horizon-f784b6cfc-t55xf\" (UID: \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\") " pod="openstack/horizon-f784b6cfc-t55xf" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.327826 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/244410df-3345-4aa9-8d98-677e5593bfad-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.327882 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/244410df-3345-4aa9-8d98-677e5593bfad-logs\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.327913 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.328032 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-config-data\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.328068 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.328089 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5zgk\" (UniqueName: \"kubernetes.io/projected/26dce0f9-35d4-47e8-8bdc-61a508050489-kube-api-access-h5zgk\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.328164 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d03bb0b9-32cd-43b8-92e4-2b8f83361246-scripts\") pod \"horizon-f784b6cfc-t55xf\" (UID: \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\") " pod="openstack/horizon-f784b6cfc-t55xf" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.328212 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh594\" (UniqueName: \"kubernetes.io/projected/d03bb0b9-32cd-43b8-92e4-2b8f83361246-kube-api-access-dh594\") pod \"horizon-f784b6cfc-t55xf\" (UID: \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\") " pod="openstack/horizon-f784b6cfc-t55xf" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.328248 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.328301 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/26dce0f9-35d4-47e8-8bdc-61a508050489-ceph\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.328329 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-config-data\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.328354 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.328408 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-scripts\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.328428 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-scripts\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.328512 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/244410df-3345-4aa9-8d98-677e5593bfad-ceph\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.328749 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/244410df-3345-4aa9-8d98-677e5593bfad-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.329025 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/244410df-3345-4aa9-8d98-677e5593bfad-logs\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.329214 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d03bb0b9-32cd-43b8-92e4-2b8f83361246-logs\") pod \"horizon-f784b6cfc-t55xf\" (UID: \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\") " pod="openstack/horizon-f784b6cfc-t55xf" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.329709 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.330514 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.330709 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.332279 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.332851 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.333023 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.335643 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/244410df-3345-4aa9-8d98-677e5593bfad-ceph\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.335854 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.341726 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.342196 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d03bb0b9-32cd-43b8-92e4-2b8f83361246-config-data\") pod \"horizon-f784b6cfc-t55xf\" (UID: \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\") " pod="openstack/horizon-f784b6cfc-t55xf" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.342771 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d03bb0b9-32cd-43b8-92e4-2b8f83361246-scripts\") pod \"horizon-f784b6cfc-t55xf\" (UID: \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\") " pod="openstack/horizon-f784b6cfc-t55xf" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.346050 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-scripts\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.346106 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh594\" (UniqueName: \"kubernetes.io/projected/d03bb0b9-32cd-43b8-92e4-2b8f83361246-kube-api-access-dh594\") pod \"horizon-f784b6cfc-t55xf\" (UID: \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\") " pod="openstack/horizon-f784b6cfc-t55xf" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.346107 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d03bb0b9-32cd-43b8-92e4-2b8f83361246-horizon-secret-key\") pod \"horizon-f784b6cfc-t55xf\" (UID: \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\") " pod="openstack/horizon-f784b6cfc-t55xf" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.347542 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.348020 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-config-data\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.349673 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf46n\" (UniqueName: \"kubernetes.io/projected/244410df-3345-4aa9-8d98-677e5593bfad-kube-api-access-hf46n\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.410617 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-psw4h" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.416343 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f784b6cfc-t55xf" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.432086 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-scripts\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.432157 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-logs\") pod \"horizon-c4458788c-rwwjv\" (UID: \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\") " pod="openstack/horizon-c4458788c-rwwjv" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.432190 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks98r\" (UniqueName: \"kubernetes.io/projected/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-kube-api-access-ks98r\") pod \"horizon-c4458788c-rwwjv\" (UID: \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\") " pod="openstack/horizon-c4458788c-rwwjv" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.432232 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-config-data\") pod \"horizon-c4458788c-rwwjv\" (UID: \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\") " pod="openstack/horizon-c4458788c-rwwjv" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.432267 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.432304 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26dce0f9-35d4-47e8-8bdc-61a508050489-logs\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.432332 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-scripts\") pod \"horizon-c4458788c-rwwjv\" (UID: \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\") " pod="openstack/horizon-c4458788c-rwwjv" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.432354 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26dce0f9-35d4-47e8-8bdc-61a508050489-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.432422 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-horizon-secret-key\") pod \"horizon-c4458788c-rwwjv\" (UID: \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\") " pod="openstack/horizon-c4458788c-rwwjv" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.432461 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-config-data\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.432475 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.432495 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5zgk\" (UniqueName: \"kubernetes.io/projected/26dce0f9-35d4-47e8-8bdc-61a508050489-kube-api-access-h5zgk\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.432549 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.432573 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/26dce0f9-35d4-47e8-8bdc-61a508050489-ceph\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.435597 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26dce0f9-35d4-47e8-8bdc-61a508050489-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.435919 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26dce0f9-35d4-47e8-8bdc-61a508050489-logs\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.438823 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.439028 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.439632 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.441600 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-scripts\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.444705 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/26dce0f9-35d4-47e8-8bdc-61a508050489-ceph\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.447315 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.453666 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-config-data\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.454397 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.458473 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.491912 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5zgk\" (UniqueName: \"kubernetes.io/projected/26dce0f9-35d4-47e8-8bdc-61a508050489-kube-api-access-h5zgk\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.506743 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.506836 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.534977 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-logs\") pod \"horizon-c4458788c-rwwjv\" (UID: \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\") " pod="openstack/horizon-c4458788c-rwwjv" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.535042 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks98r\" (UniqueName: \"kubernetes.io/projected/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-kube-api-access-ks98r\") pod \"horizon-c4458788c-rwwjv\" (UID: \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\") " pod="openstack/horizon-c4458788c-rwwjv" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.535081 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-config-data\") pod \"horizon-c4458788c-rwwjv\" (UID: \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\") " pod="openstack/horizon-c4458788c-rwwjv" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.535143 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-scripts\") pod \"horizon-c4458788c-rwwjv\" (UID: \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\") " pod="openstack/horizon-c4458788c-rwwjv" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.535215 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-horizon-secret-key\") pod \"horizon-c4458788c-rwwjv\" (UID: \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\") " pod="openstack/horizon-c4458788c-rwwjv" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.536869 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-logs\") pod \"horizon-c4458788c-rwwjv\" (UID: \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\") " pod="openstack/horizon-c4458788c-rwwjv" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.538139 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-scripts\") pod \"horizon-c4458788c-rwwjv\" (UID: \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\") " pod="openstack/horizon-c4458788c-rwwjv" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.539738 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-config-data\") pod \"horizon-c4458788c-rwwjv\" (UID: \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\") " pod="openstack/horizon-c4458788c-rwwjv" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.552768 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-horizon-secret-key\") pod \"horizon-c4458788c-rwwjv\" (UID: \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\") " pod="openstack/horizon-c4458788c-rwwjv" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.555411 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks98r\" (UniqueName: \"kubernetes.io/projected/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-kube-api-access-ks98r\") pod \"horizon-c4458788c-rwwjv\" (UID: \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\") " pod="openstack/horizon-c4458788c-rwwjv" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.588783 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c4458788c-rwwjv" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.658471 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-v2bv5"] Nov 22 05:45:28 crc kubenswrapper[4818]: W1122 05:45:28.748533 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod384fe684_9694_4069_b37e_ad50f83ef867.slice/crio-2b0c6cb84644351b03e0ca91345972e188f24dc05df432b5ff29d63152a7e6a5 WatchSource:0}: Error finding container 2b0c6cb84644351b03e0ca91345972e188f24dc05df432b5ff29d63152a7e6a5: Status 404 returned error can't find the container with id 2b0c6cb84644351b03e0ca91345972e188f24dc05df432b5ff29d63152a7e6a5 Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.757191 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-50ec-account-create-zq678"] Nov 22 05:45:28 crc kubenswrapper[4818]: W1122 05:45:28.763818 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ca6675b_db38_49ba_a7ab_396d5d6dc54f.slice/crio-7a8d95cb62aadaf880a097a7800cbc0a387b5830910d3b5872869c297fa954bb WatchSource:0}: Error finding container 7a8d95cb62aadaf880a097a7800cbc0a387b5830910d3b5872869c297fa954bb: Status 404 returned error can't find the container with id 7a8d95cb62aadaf880a097a7800cbc0a387b5830910d3b5872869c297fa954bb Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.764042 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"0c1e7078-787f-4725-80b1-45f7047e108f","Type":"ContainerStarted","Data":"d3874de15447d9744108a937874829fde38f1dab3f37e39e83a60410876ba25d"} Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.766429 4818 generic.go:334] "Generic (PLEG): container finished" podID="e3759d20-d679-42de-9a4e-f32305ec1ca0" containerID="aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514" exitCode=0 Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.766504 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ksfkm" event={"ID":"e3759d20-d679-42de-9a4e-f32305ec1ca0","Type":"ContainerDied","Data":"aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514"} Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.766532 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ksfkm" event={"ID":"e3759d20-d679-42de-9a4e-f32305ec1ca0","Type":"ContainerStarted","Data":"111f708a9cacde7583bf2e26ada8a9d1460f6d13ff5ffd9b987c416bf415431a"} Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.770432 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.770994 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"3dfef980-4eab-42b9-821e-233a264843cf","Type":"ContainerStarted","Data":"5853d6e52ab97c7fea393b6d5923ccffeba959644236a8b2b3a6a00108baa854"} Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.771051 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.784672 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.791700 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.797163 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.944194 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/244410df-3345-4aa9-8d98-677e5593bfad-logs\") pod \"244410df-3345-4aa9-8d98-677e5593bfad\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.944368 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-config-data\") pod \"26dce0f9-35d4-47e8-8bdc-61a508050489\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.944461 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/244410df-3345-4aa9-8d98-677e5593bfad-logs" (OuterVolumeSpecName: "logs") pod "244410df-3345-4aa9-8d98-677e5593bfad" (UID: "244410df-3345-4aa9-8d98-677e5593bfad"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.944394 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"26dce0f9-35d4-47e8-8bdc-61a508050489\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.948521 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-config-data\") pod \"244410df-3345-4aa9-8d98-677e5593bfad\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.948574 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-combined-ca-bundle\") pod \"26dce0f9-35d4-47e8-8bdc-61a508050489\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.948607 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-scripts\") pod \"244410df-3345-4aa9-8d98-677e5593bfad\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.948651 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-public-tls-certs\") pod \"244410df-3345-4aa9-8d98-677e5593bfad\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.948671 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/26dce0f9-35d4-47e8-8bdc-61a508050489-ceph\") pod \"26dce0f9-35d4-47e8-8bdc-61a508050489\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.948722 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26dce0f9-35d4-47e8-8bdc-61a508050489-httpd-run\") pod \"26dce0f9-35d4-47e8-8bdc-61a508050489\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.948756 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5zgk\" (UniqueName: \"kubernetes.io/projected/26dce0f9-35d4-47e8-8bdc-61a508050489-kube-api-access-h5zgk\") pod \"26dce0f9-35d4-47e8-8bdc-61a508050489\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.948807 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-internal-tls-certs\") pod \"26dce0f9-35d4-47e8-8bdc-61a508050489\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.948860 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-combined-ca-bundle\") pod \"244410df-3345-4aa9-8d98-677e5593bfad\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.948895 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hf46n\" (UniqueName: \"kubernetes.io/projected/244410df-3345-4aa9-8d98-677e5593bfad-kube-api-access-hf46n\") pod \"244410df-3345-4aa9-8d98-677e5593bfad\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.948926 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/244410df-3345-4aa9-8d98-677e5593bfad-httpd-run\") pod \"244410df-3345-4aa9-8d98-677e5593bfad\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.948951 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/244410df-3345-4aa9-8d98-677e5593bfad-ceph\") pod \"244410df-3345-4aa9-8d98-677e5593bfad\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.949058 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26dce0f9-35d4-47e8-8bdc-61a508050489-logs\") pod \"26dce0f9-35d4-47e8-8bdc-61a508050489\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.949122 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-scripts\") pod \"26dce0f9-35d4-47e8-8bdc-61a508050489\" (UID: \"26dce0f9-35d4-47e8-8bdc-61a508050489\") " Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.949149 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"244410df-3345-4aa9-8d98-677e5593bfad\" (UID: \"244410df-3345-4aa9-8d98-677e5593bfad\") " Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.949900 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/244410df-3345-4aa9-8d98-677e5593bfad-logs\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.950869 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26dce0f9-35d4-47e8-8bdc-61a508050489-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "26dce0f9-35d4-47e8-8bdc-61a508050489" (UID: "26dce0f9-35d4-47e8-8bdc-61a508050489"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.951063 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26dce0f9-35d4-47e8-8bdc-61a508050489-logs" (OuterVolumeSpecName: "logs") pod "26dce0f9-35d4-47e8-8bdc-61a508050489" (UID: "26dce0f9-35d4-47e8-8bdc-61a508050489"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.951247 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/244410df-3345-4aa9-8d98-677e5593bfad-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "244410df-3345-4aa9-8d98-677e5593bfad" (UID: "244410df-3345-4aa9-8d98-677e5593bfad"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.952117 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "26dce0f9-35d4-47e8-8bdc-61a508050489" (UID: "26dce0f9-35d4-47e8-8bdc-61a508050489"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.956356 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-config-data" (OuterVolumeSpecName: "config-data") pod "244410df-3345-4aa9-8d98-677e5593bfad" (UID: "244410df-3345-4aa9-8d98-677e5593bfad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.956372 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26dce0f9-35d4-47e8-8bdc-61a508050489" (UID: "26dce0f9-35d4-47e8-8bdc-61a508050489"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.956370 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "244410df-3345-4aa9-8d98-677e5593bfad" (UID: "244410df-3345-4aa9-8d98-677e5593bfad"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.956634 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f784b6cfc-t55xf"] Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.959592 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-config-data" (OuterVolumeSpecName: "config-data") pod "26dce0f9-35d4-47e8-8bdc-61a508050489" (UID: "26dce0f9-35d4-47e8-8bdc-61a508050489"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.959780 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "244410df-3345-4aa9-8d98-677e5593bfad" (UID: "244410df-3345-4aa9-8d98-677e5593bfad"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.959768 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-scripts" (OuterVolumeSpecName: "scripts") pod "244410df-3345-4aa9-8d98-677e5593bfad" (UID: "244410df-3345-4aa9-8d98-677e5593bfad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.960081 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-scripts" (OuterVolumeSpecName: "scripts") pod "26dce0f9-35d4-47e8-8bdc-61a508050489" (UID: "26dce0f9-35d4-47e8-8bdc-61a508050489"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.960303 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/244410df-3345-4aa9-8d98-677e5593bfad-ceph" (OuterVolumeSpecName: "ceph") pod "244410df-3345-4aa9-8d98-677e5593bfad" (UID: "244410df-3345-4aa9-8d98-677e5593bfad"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.960361 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26dce0f9-35d4-47e8-8bdc-61a508050489-ceph" (OuterVolumeSpecName: "ceph") pod "26dce0f9-35d4-47e8-8bdc-61a508050489" (UID: "26dce0f9-35d4-47e8-8bdc-61a508050489"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.961411 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "26dce0f9-35d4-47e8-8bdc-61a508050489" (UID: "26dce0f9-35d4-47e8-8bdc-61a508050489"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.967763 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "244410df-3345-4aa9-8d98-677e5593bfad" (UID: "244410df-3345-4aa9-8d98-677e5593bfad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.967785 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26dce0f9-35d4-47e8-8bdc-61a508050489-kube-api-access-h5zgk" (OuterVolumeSpecName: "kube-api-access-h5zgk") pod "26dce0f9-35d4-47e8-8bdc-61a508050489" (UID: "26dce0f9-35d4-47e8-8bdc-61a508050489"). InnerVolumeSpecName "kube-api-access-h5zgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:45:28 crc kubenswrapper[4818]: I1122 05:45:28.969344 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/244410df-3345-4aa9-8d98-677e5593bfad-kube-api-access-hf46n" (OuterVolumeSpecName: "kube-api-access-hf46n") pod "244410df-3345-4aa9-8d98-677e5593bfad" (UID: "244410df-3345-4aa9-8d98-677e5593bfad"). InnerVolumeSpecName "kube-api-access-hf46n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.051620 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26dce0f9-35d4-47e8-8bdc-61a508050489-logs\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.051652 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.051684 4818 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.051695 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.051709 4818 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.051717 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.051726 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.051735 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.051746 4818 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.051754 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/26dce0f9-35d4-47e8-8bdc-61a508050489-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.051764 4818 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26dce0f9-35d4-47e8-8bdc-61a508050489-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.051774 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5zgk\" (UniqueName: \"kubernetes.io/projected/26dce0f9-35d4-47e8-8bdc-61a508050489-kube-api-access-h5zgk\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.051784 4818 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26dce0f9-35d4-47e8-8bdc-61a508050489-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.051795 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/244410df-3345-4aa9-8d98-677e5593bfad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.051806 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hf46n\" (UniqueName: \"kubernetes.io/projected/244410df-3345-4aa9-8d98-677e5593bfad-kube-api-access-hf46n\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.051816 4818 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/244410df-3345-4aa9-8d98-677e5593bfad-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.051826 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/244410df-3345-4aa9-8d98-677e5593bfad-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.074909 4818 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.076586 4818 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.103578 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c4458788c-rwwjv"] Nov 22 05:45:29 crc kubenswrapper[4818]: W1122 05:45:29.123684 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90e9c626_fc6c_4ad9_b385_e634bd3dbfb8.slice/crio-e1eab19f7fcc992ae62d5b3e24b94306abc74ef192d1320e51fd787317edb8ac WatchSource:0}: Error finding container e1eab19f7fcc992ae62d5b3e24b94306abc74ef192d1320e51fd787317edb8ac: Status 404 returned error can't find the container with id e1eab19f7fcc992ae62d5b3e24b94306abc74ef192d1320e51fd787317edb8ac Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.153278 4818 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.153307 4818 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.782893 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c4458788c-rwwjv" event={"ID":"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8","Type":"ContainerStarted","Data":"e1eab19f7fcc992ae62d5b3e24b94306abc74ef192d1320e51fd787317edb8ac"} Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.787991 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"0c1e7078-787f-4725-80b1-45f7047e108f","Type":"ContainerStarted","Data":"e5535d36a5231f66f7b0699d1b3692c8d35d38131877d7f3f7ab8f3d7bed9a7b"} Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.788199 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"0c1e7078-787f-4725-80b1-45f7047e108f","Type":"ContainerStarted","Data":"e9f5587393be90812592b4bc0271562dd6eaf2617c4027d60079f6866903ef91"} Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.801935 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ksfkm" event={"ID":"e3759d20-d679-42de-9a4e-f32305ec1ca0","Type":"ContainerStarted","Data":"f41b189f96fd2137bd4914485c32775c0dfb95bb75ea890ad134e6edfdc260b4"} Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.805265 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"3dfef980-4eab-42b9-821e-233a264843cf","Type":"ContainerStarted","Data":"8c572d75c540da7431b480300d87d7f48414721621e6ad9aebaafbbd2fb979e6"} Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.808672 4818 generic.go:334] "Generic (PLEG): container finished" podID="384fe684-9694-4069-b37e-ad50f83ef867" containerID="a9f00296718de0e7ebca022ccc1ce4981fe7d40d35576e1abf7278830f38ceb8" exitCode=0 Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.808762 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-v2bv5" event={"ID":"384fe684-9694-4069-b37e-ad50f83ef867","Type":"ContainerDied","Data":"a9f00296718de0e7ebca022ccc1ce4981fe7d40d35576e1abf7278830f38ceb8"} Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.808789 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-v2bv5" event={"ID":"384fe684-9694-4069-b37e-ad50f83ef867","Type":"ContainerStarted","Data":"2b0c6cb84644351b03e0ca91345972e188f24dc05df432b5ff29d63152a7e6a5"} Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.810879 4818 generic.go:334] "Generic (PLEG): container finished" podID="1ca6675b-db38-49ba-a7ab-396d5d6dc54f" containerID="1bc8c9b47327022b405fcdcb1246ffd08cae2061a2c32c8301a924e45630d081" exitCode=0 Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.810929 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-50ec-account-create-zq678" event={"ID":"1ca6675b-db38-49ba-a7ab-396d5d6dc54f","Type":"ContainerDied","Data":"1bc8c9b47327022b405fcdcb1246ffd08cae2061a2c32c8301a924e45630d081"} Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.810951 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-50ec-account-create-zq678" event={"ID":"1ca6675b-db38-49ba-a7ab-396d5d6dc54f","Type":"ContainerStarted","Data":"7a8d95cb62aadaf880a097a7800cbc0a387b5830910d3b5872869c297fa954bb"} Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.813209 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.817270 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.818788 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f784b6cfc-t55xf" event={"ID":"d03bb0b9-32cd-43b8-92e4-2b8f83361246","Type":"ContainerStarted","Data":"46ebc9a94338c8ec0c1da411417b3b80dc639867a75e00d379558ce3aaeedc22"} Nov 22 05:45:29 crc kubenswrapper[4818]: I1122 05:45:29.833500 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.189219293 podStartE2EDuration="2.833484147s" podCreationTimestamp="2025-11-22 05:45:27 +0000 UTC" firstStartedPulling="2025-11-22 05:45:28.147413701 +0000 UTC m=+3480.721830228" lastFinishedPulling="2025-11-22 05:45:28.791678545 +0000 UTC m=+3481.366095082" observedRunningTime="2025-11-22 05:45:29.819147055 +0000 UTC m=+3482.393563582" watchObservedRunningTime="2025-11-22 05:45:29.833484147 +0000 UTC m=+3482.407900674" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.028462 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.036371 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.063134 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.067729 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.071112 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.071534 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.072108 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2zbds" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.090730 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.093290 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.106716 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.117042 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.130897 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.132871 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.138027 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.138232 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.145149 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.195700 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.195852 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.195894 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf03a664-65df-4df2-a074-d35397ff3913-logs\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.195911 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/bf03a664-65df-4df2-a074-d35397ff3913-ceph\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.195959 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn2wh\" (UniqueName: \"kubernetes.io/projected/bf03a664-65df-4df2-a074-d35397ff3913-kube-api-access-wn2wh\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.195992 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-config-data\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.196013 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf03a664-65df-4df2-a074-d35397ff3913-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.196043 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.196067 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-scripts\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.297189 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.297473 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn2wh\" (UniqueName: \"kubernetes.io/projected/bf03a664-65df-4df2-a074-d35397ff3913-kube-api-access-wn2wh\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.297508 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3a4ab9f9-573b-4953-8683-1940e6678c5f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.297529 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-config-data\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.297544 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3a4ab9f9-573b-4953-8683-1940e6678c5f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.297566 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.297587 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvjcf\" (UniqueName: \"kubernetes.io/projected/3a4ab9f9-573b-4953-8683-1940e6678c5f-kube-api-access-xvjcf\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.297609 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf03a664-65df-4df2-a074-d35397ff3913-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.297632 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a4ab9f9-573b-4953-8683-1940e6678c5f-logs\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.297655 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.297678 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-scripts\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.297699 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.297728 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.297745 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.297775 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.297804 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.297828 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf03a664-65df-4df2-a074-d35397ff3913-logs\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.297846 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/bf03a664-65df-4df2-a074-d35397ff3913-ceph\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.298876 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf03a664-65df-4df2-a074-d35397ff3913-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.301689 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf03a664-65df-4df2-a074-d35397ff3913-logs\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.301776 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.307917 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.317451 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-scripts\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.317731 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/bf03a664-65df-4df2-a074-d35397ff3913-ceph\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.334378 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="244410df-3345-4aa9-8d98-677e5593bfad" path="/var/lib/kubelet/pods/244410df-3345-4aa9-8d98-677e5593bfad/volumes" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.345793 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-config-data\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.346382 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.346562 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26dce0f9-35d4-47e8-8bdc-61a508050489" path="/var/lib/kubelet/pods/26dce0f9-35d4-47e8-8bdc-61a508050489/volumes" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.356985 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn2wh\" (UniqueName: \"kubernetes.io/projected/bf03a664-65df-4df2-a074-d35397ff3913-kube-api-access-wn2wh\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.399762 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.399798 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.399900 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.399959 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3a4ab9f9-573b-4953-8683-1940e6678c5f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.399977 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3a4ab9f9-573b-4953-8683-1940e6678c5f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.399990 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.400008 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvjcf\" (UniqueName: \"kubernetes.io/projected/3a4ab9f9-573b-4953-8683-1940e6678c5f-kube-api-access-xvjcf\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.400040 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a4ab9f9-573b-4953-8683-1940e6678c5f-logs\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.400085 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.404936 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.414442 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.415497 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a4ab9f9-573b-4953-8683-1940e6678c5f-logs\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.415695 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3a4ab9f9-573b-4953-8683-1940e6678c5f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.420319 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.434236 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3a4ab9f9-573b-4953-8683-1940e6678c5f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.436664 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.437448 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.439522 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.464342 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.470612 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvjcf\" (UniqueName: \"kubernetes.io/projected/3a4ab9f9-573b-4953-8683-1940e6678c5f-kube-api-access-xvjcf\") pod \"glance-default-internal-api-0\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.471659 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.702980 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.733150 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-f784b6cfc-t55xf"] Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.758858 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7764859bb4-tnzkk"] Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.762617 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.771148 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.779850 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7764859bb4-tnzkk"] Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.790594 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.864041 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c4458788c-rwwjv"] Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.864415 4818 generic.go:334] "Generic (PLEG): container finished" podID="e3759d20-d679-42de-9a4e-f32305ec1ca0" containerID="f41b189f96fd2137bd4914485c32775c0dfb95bb75ea890ad134e6edfdc260b4" exitCode=0 Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.864533 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ksfkm" event={"ID":"e3759d20-d679-42de-9a4e-f32305ec1ca0","Type":"ContainerDied","Data":"f41b189f96fd2137bd4914485c32775c0dfb95bb75ea890ad134e6edfdc260b4"} Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.888900 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"3dfef980-4eab-42b9-821e-233a264843cf","Type":"ContainerStarted","Data":"90cd0d89e53328d3e2a84dcb0e404f485f33302074bb82d0e5d8eb3ee9d8f607"} Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.894032 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.912072 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-config-data\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.912130 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-horizon-tls-certs\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.912187 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-scripts\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.912203 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tsmx\" (UniqueName: \"kubernetes.io/projected/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-kube-api-access-7tsmx\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.912227 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-horizon-secret-key\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.912276 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-combined-ca-bundle\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.912301 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-logs\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.914312 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-86757bfcb-s9bbk"] Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.915898 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.936464 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-86757bfcb-s9bbk"] Nov 22 05:45:30 crc kubenswrapper[4818]: I1122 05:45:30.962523 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=3.303054625 podStartE2EDuration="3.962505734s" podCreationTimestamp="2025-11-22 05:45:27 +0000 UTC" firstStartedPulling="2025-11-22 05:45:28.509414131 +0000 UTC m=+3481.083830648" lastFinishedPulling="2025-11-22 05:45:29.16886523 +0000 UTC m=+3481.743281757" observedRunningTime="2025-11-22 05:45:30.925429345 +0000 UTC m=+3483.499845882" watchObservedRunningTime="2025-11-22 05:45:30.962505734 +0000 UTC m=+3483.536922261" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.013850 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34f2acb8-906b-4520-86c2-bb603a964489-combined-ca-bundle\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.014134 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/34f2acb8-906b-4520-86c2-bb603a964489-horizon-secret-key\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.014172 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-config-data\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.014312 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-horizon-tls-certs\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.014335 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34f2acb8-906b-4520-86c2-bb603a964489-scripts\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.014353 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/34f2acb8-906b-4520-86c2-bb603a964489-horizon-tls-certs\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.014395 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/34f2acb8-906b-4520-86c2-bb603a964489-config-data\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.014431 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34f2acb8-906b-4520-86c2-bb603a964489-logs\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.014448 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cps68\" (UniqueName: \"kubernetes.io/projected/34f2acb8-906b-4520-86c2-bb603a964489-kube-api-access-cps68\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.014475 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-scripts\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.014490 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tsmx\" (UniqueName: \"kubernetes.io/projected/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-kube-api-access-7tsmx\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.014514 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-horizon-secret-key\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.014571 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-combined-ca-bundle\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.014601 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-logs\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.015841 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-scripts\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.017770 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-logs\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.018885 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-config-data\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.023424 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-horizon-tls-certs\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.027287 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-horizon-secret-key\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.031981 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-combined-ca-bundle\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.040770 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tsmx\" (UniqueName: \"kubernetes.io/projected/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-kube-api-access-7tsmx\") pod \"horizon-7764859bb4-tnzkk\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.100343 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.118706 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34f2acb8-906b-4520-86c2-bb603a964489-combined-ca-bundle\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.119675 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/34f2acb8-906b-4520-86c2-bb603a964489-horizon-secret-key\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.119777 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34f2acb8-906b-4520-86c2-bb603a964489-scripts\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.119802 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/34f2acb8-906b-4520-86c2-bb603a964489-horizon-tls-certs\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.119871 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/34f2acb8-906b-4520-86c2-bb603a964489-config-data\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.119900 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34f2acb8-906b-4520-86c2-bb603a964489-logs\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.119924 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cps68\" (UniqueName: \"kubernetes.io/projected/34f2acb8-906b-4520-86c2-bb603a964489-kube-api-access-cps68\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.120542 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34f2acb8-906b-4520-86c2-bb603a964489-scripts\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.120793 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34f2acb8-906b-4520-86c2-bb603a964489-logs\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.121013 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/34f2acb8-906b-4520-86c2-bb603a964489-config-data\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.123943 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/34f2acb8-906b-4520-86c2-bb603a964489-horizon-secret-key\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.124589 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/34f2acb8-906b-4520-86c2-bb603a964489-horizon-tls-certs\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.127076 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34f2acb8-906b-4520-86c2-bb603a964489-combined-ca-bundle\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.140983 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cps68\" (UniqueName: \"kubernetes.io/projected/34f2acb8-906b-4520-86c2-bb603a964489-kube-api-access-cps68\") pod \"horizon-86757bfcb-s9bbk\" (UID: \"34f2acb8-906b-4520-86c2-bb603a964489\") " pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.173068 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.235496 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.486727 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-v2bv5" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.606590 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-50ec-account-create-zq678" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.666628 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfh7r\" (UniqueName: \"kubernetes.io/projected/384fe684-9694-4069-b37e-ad50f83ef867-kube-api-access-dfh7r\") pod \"384fe684-9694-4069-b37e-ad50f83ef867\" (UID: \"384fe684-9694-4069-b37e-ad50f83ef867\") " Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.666803 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/384fe684-9694-4069-b37e-ad50f83ef867-operator-scripts\") pod \"384fe684-9694-4069-b37e-ad50f83ef867\" (UID: \"384fe684-9694-4069-b37e-ad50f83ef867\") " Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.667455 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/384fe684-9694-4069-b37e-ad50f83ef867-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "384fe684-9694-4069-b37e-ad50f83ef867" (UID: "384fe684-9694-4069-b37e-ad50f83ef867"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.678308 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/384fe684-9694-4069-b37e-ad50f83ef867-kube-api-access-dfh7r" (OuterVolumeSpecName: "kube-api-access-dfh7r") pod "384fe684-9694-4069-b37e-ad50f83ef867" (UID: "384fe684-9694-4069-b37e-ad50f83ef867"). InnerVolumeSpecName "kube-api-access-dfh7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.706810 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7764859bb4-tnzkk"] Nov 22 05:45:31 crc kubenswrapper[4818]: W1122 05:45:31.716474 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60b9da36_1281_4cc5_a9ac_a2df2e4d9638.slice/crio-7c11e3c6df9c02580c092fb06c3874060f2efafbabbc374db2914c8fd2166d13 WatchSource:0}: Error finding container 7c11e3c6df9c02580c092fb06c3874060f2efafbabbc374db2914c8fd2166d13: Status 404 returned error can't find the container with id 7c11e3c6df9c02580c092fb06c3874060f2efafbabbc374db2914c8fd2166d13 Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.717442 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.768110 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bds4t\" (UniqueName: \"kubernetes.io/projected/1ca6675b-db38-49ba-a7ab-396d5d6dc54f-kube-api-access-bds4t\") pod \"1ca6675b-db38-49ba-a7ab-396d5d6dc54f\" (UID: \"1ca6675b-db38-49ba-a7ab-396d5d6dc54f\") " Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.769160 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ca6675b-db38-49ba-a7ab-396d5d6dc54f-operator-scripts\") pod \"1ca6675b-db38-49ba-a7ab-396d5d6dc54f\" (UID: \"1ca6675b-db38-49ba-a7ab-396d5d6dc54f\") " Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.770421 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfh7r\" (UniqueName: \"kubernetes.io/projected/384fe684-9694-4069-b37e-ad50f83ef867-kube-api-access-dfh7r\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.770441 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/384fe684-9694-4069-b37e-ad50f83ef867-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.770954 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ca6675b-db38-49ba-a7ab-396d5d6dc54f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1ca6675b-db38-49ba-a7ab-396d5d6dc54f" (UID: "1ca6675b-db38-49ba-a7ab-396d5d6dc54f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.779825 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ca6675b-db38-49ba-a7ab-396d5d6dc54f-kube-api-access-bds4t" (OuterVolumeSpecName: "kube-api-access-bds4t") pod "1ca6675b-db38-49ba-a7ab-396d5d6dc54f" (UID: "1ca6675b-db38-49ba-a7ab-396d5d6dc54f"). InnerVolumeSpecName "kube-api-access-bds4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:45:31 crc kubenswrapper[4818]: E1122 05:45:31.780328 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3759d20_d679_42de_9a4e_f32305ec1ca0.slice/crio-conmon-aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3759d20_d679_42de_9a4e_f32305ec1ca0.slice/crio-aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514.scope\": RecentStats: unable to find data in memory cache]" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.872880 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bds4t\" (UniqueName: \"kubernetes.io/projected/1ca6675b-db38-49ba-a7ab-396d5d6dc54f-kube-api-access-bds4t\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.873155 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ca6675b-db38-49ba-a7ab-396d5d6dc54f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.909671 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3a4ab9f9-573b-4953-8683-1940e6678c5f","Type":"ContainerStarted","Data":"267ac408cba6aff5449637b3c3e47043f9452c11e2d06dff1f51b5dd093b4b2f"} Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.921734 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ksfkm" event={"ID":"e3759d20-d679-42de-9a4e-f32305ec1ca0","Type":"ContainerStarted","Data":"e499a2f0e3524f430fa8bb42729769b665dc5cb1795195e03f7620c112479094"} Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.924434 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7764859bb4-tnzkk" event={"ID":"60b9da36-1281-4cc5-a9ac-a2df2e4d9638","Type":"ContainerStarted","Data":"7c11e3c6df9c02580c092fb06c3874060f2efafbabbc374db2914c8fd2166d13"} Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.926240 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bf03a664-65df-4df2-a074-d35397ff3913","Type":"ContainerStarted","Data":"788cede5865c6c22804ff1125f51dff431a47b52687dd24668a1b35d6affde47"} Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.932612 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-v2bv5" event={"ID":"384fe684-9694-4069-b37e-ad50f83ef867","Type":"ContainerDied","Data":"2b0c6cb84644351b03e0ca91345972e188f24dc05df432b5ff29d63152a7e6a5"} Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.932649 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b0c6cb84644351b03e0ca91345972e188f24dc05df432b5ff29d63152a7e6a5" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.932694 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-v2bv5" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.947000 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-50ec-account-create-zq678" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.947870 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ksfkm" podStartSLOduration=2.241639389 podStartE2EDuration="4.94785219s" podCreationTimestamp="2025-11-22 05:45:27 +0000 UTC" firstStartedPulling="2025-11-22 05:45:28.789961259 +0000 UTC m=+3481.364377786" lastFinishedPulling="2025-11-22 05:45:31.49617406 +0000 UTC m=+3484.070590587" observedRunningTime="2025-11-22 05:45:31.938179733 +0000 UTC m=+3484.512596260" watchObservedRunningTime="2025-11-22 05:45:31.94785219 +0000 UTC m=+3484.522268717" Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.948730 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-50ec-account-create-zq678" event={"ID":"1ca6675b-db38-49ba-a7ab-396d5d6dc54f","Type":"ContainerDied","Data":"7a8d95cb62aadaf880a097a7800cbc0a387b5830910d3b5872869c297fa954bb"} Nov 22 05:45:31 crc kubenswrapper[4818]: I1122 05:45:31.948762 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a8d95cb62aadaf880a097a7800cbc0a387b5830910d3b5872869c297fa954bb" Nov 22 05:45:32 crc kubenswrapper[4818]: I1122 05:45:32.029927 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-86757bfcb-s9bbk"] Nov 22 05:45:32 crc kubenswrapper[4818]: I1122 05:45:32.517666 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Nov 22 05:45:32 crc kubenswrapper[4818]: I1122 05:45:32.697512 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:32 crc kubenswrapper[4818]: I1122 05:45:32.961381 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86757bfcb-s9bbk" event={"ID":"34f2acb8-906b-4520-86c2-bb603a964489","Type":"ContainerStarted","Data":"054d4814c20c250af0e5a3c04ddd4553d8b919cbf48ecbe83a1c65cb039b6ae7"} Nov 22 05:45:32 crc kubenswrapper[4818]: I1122 05:45:32.963355 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bf03a664-65df-4df2-a074-d35397ff3913","Type":"ContainerStarted","Data":"78683020f40cca0195bb98a680aa641ab7b2b9ed9a3b815213f6f1c7af2909fc"} Nov 22 05:45:32 crc kubenswrapper[4818]: I1122 05:45:32.965391 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3a4ab9f9-573b-4953-8683-1940e6678c5f","Type":"ContainerStarted","Data":"f90bd608574b9d6050a979456443e31d4dcd782f7b71be215290a26a30506b74"} Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.249220 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-2xttz"] Nov 22 05:45:33 crc kubenswrapper[4818]: E1122 05:45:33.250088 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="384fe684-9694-4069-b37e-ad50f83ef867" containerName="mariadb-database-create" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.250109 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="384fe684-9694-4069-b37e-ad50f83ef867" containerName="mariadb-database-create" Nov 22 05:45:33 crc kubenswrapper[4818]: E1122 05:45:33.250143 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca6675b-db38-49ba-a7ab-396d5d6dc54f" containerName="mariadb-account-create" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.250150 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca6675b-db38-49ba-a7ab-396d5d6dc54f" containerName="mariadb-account-create" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.250336 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="384fe684-9694-4069-b37e-ad50f83ef867" containerName="mariadb-database-create" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.250350 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca6675b-db38-49ba-a7ab-396d5d6dc54f" containerName="mariadb-account-create" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.250994 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-2xttz" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.253365 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-5jfbb" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.253725 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.273663 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-2xttz"] Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.408753 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d58gq\" (UniqueName: \"kubernetes.io/projected/64f4e98d-2953-472c-852a-424b72c513a9-kube-api-access-d58gq\") pod \"manila-db-sync-2xttz\" (UID: \"64f4e98d-2953-472c-852a-424b72c513a9\") " pod="openstack/manila-db-sync-2xttz" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.408994 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64f4e98d-2953-472c-852a-424b72c513a9-combined-ca-bundle\") pod \"manila-db-sync-2xttz\" (UID: \"64f4e98d-2953-472c-852a-424b72c513a9\") " pod="openstack/manila-db-sync-2xttz" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.409016 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/64f4e98d-2953-472c-852a-424b72c513a9-job-config-data\") pod \"manila-db-sync-2xttz\" (UID: \"64f4e98d-2953-472c-852a-424b72c513a9\") " pod="openstack/manila-db-sync-2xttz" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.409101 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64f4e98d-2953-472c-852a-424b72c513a9-config-data\") pod \"manila-db-sync-2xttz\" (UID: \"64f4e98d-2953-472c-852a-424b72c513a9\") " pod="openstack/manila-db-sync-2xttz" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.511357 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64f4e98d-2953-472c-852a-424b72c513a9-config-data\") pod \"manila-db-sync-2xttz\" (UID: \"64f4e98d-2953-472c-852a-424b72c513a9\") " pod="openstack/manila-db-sync-2xttz" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.512127 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d58gq\" (UniqueName: \"kubernetes.io/projected/64f4e98d-2953-472c-852a-424b72c513a9-kube-api-access-d58gq\") pod \"manila-db-sync-2xttz\" (UID: \"64f4e98d-2953-472c-852a-424b72c513a9\") " pod="openstack/manila-db-sync-2xttz" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.512208 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64f4e98d-2953-472c-852a-424b72c513a9-combined-ca-bundle\") pod \"manila-db-sync-2xttz\" (UID: \"64f4e98d-2953-472c-852a-424b72c513a9\") " pod="openstack/manila-db-sync-2xttz" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.512227 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/64f4e98d-2953-472c-852a-424b72c513a9-job-config-data\") pod \"manila-db-sync-2xttz\" (UID: \"64f4e98d-2953-472c-852a-424b72c513a9\") " pod="openstack/manila-db-sync-2xttz" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.519095 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64f4e98d-2953-472c-852a-424b72c513a9-combined-ca-bundle\") pod \"manila-db-sync-2xttz\" (UID: \"64f4e98d-2953-472c-852a-424b72c513a9\") " pod="openstack/manila-db-sync-2xttz" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.525480 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64f4e98d-2953-472c-852a-424b72c513a9-config-data\") pod \"manila-db-sync-2xttz\" (UID: \"64f4e98d-2953-472c-852a-424b72c513a9\") " pod="openstack/manila-db-sync-2xttz" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.525833 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/64f4e98d-2953-472c-852a-424b72c513a9-job-config-data\") pod \"manila-db-sync-2xttz\" (UID: \"64f4e98d-2953-472c-852a-424b72c513a9\") " pod="openstack/manila-db-sync-2xttz" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.527708 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d58gq\" (UniqueName: \"kubernetes.io/projected/64f4e98d-2953-472c-852a-424b72c513a9-kube-api-access-d58gq\") pod \"manila-db-sync-2xttz\" (UID: \"64f4e98d-2953-472c-852a-424b72c513a9\") " pod="openstack/manila-db-sync-2xttz" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.594015 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-2xttz" Nov 22 05:45:33 crc kubenswrapper[4818]: I1122 05:45:33.988051 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-2xttz"] Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.002536 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3a4ab9f9-573b-4953-8683-1940e6678c5f","Type":"ContainerStarted","Data":"9298b0c1209abe09e04105d98507048a99335a51312bda567ab76751b8da4c9c"} Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.002864 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="3a4ab9f9-573b-4953-8683-1940e6678c5f" containerName="glance-log" containerID="cri-o://f90bd608574b9d6050a979456443e31d4dcd782f7b71be215290a26a30506b74" gracePeriod=30 Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.003205 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="3a4ab9f9-573b-4953-8683-1940e6678c5f" containerName="glance-httpd" containerID="cri-o://9298b0c1209abe09e04105d98507048a99335a51312bda567ab76751b8da4c9c" gracePeriod=30 Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.008767 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bf03a664-65df-4df2-a074-d35397ff3913","Type":"ContainerStarted","Data":"e0891fe59ab42617392296a9421c8e855213d8a70c68190080712cd69e5b56ce"} Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.008907 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="bf03a664-65df-4df2-a074-d35397ff3913" containerName="glance-log" containerID="cri-o://78683020f40cca0195bb98a680aa641ab7b2b9ed9a3b815213f6f1c7af2909fc" gracePeriod=30 Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.008974 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="bf03a664-65df-4df2-a074-d35397ff3913" containerName="glance-httpd" containerID="cri-o://e0891fe59ab42617392296a9421c8e855213d8a70c68190080712cd69e5b56ce" gracePeriod=30 Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.044064 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.04404731 podStartE2EDuration="5.04404731s" podCreationTimestamp="2025-11-22 05:45:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:45:34.04256752 +0000 UTC m=+3486.616984037" watchObservedRunningTime="2025-11-22 05:45:34.04404731 +0000 UTC m=+3486.618463837" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.047768 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.047749838 podStartE2EDuration="5.047749838s" podCreationTimestamp="2025-11-22 05:45:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:45:34.023564503 +0000 UTC m=+3486.597981030" watchObservedRunningTime="2025-11-22 05:45:34.047749838 +0000 UTC m=+3486.622166365" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.757914 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.760440 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"3a4ab9f9-573b-4953-8683-1940e6678c5f\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.760500 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3a4ab9f9-573b-4953-8683-1940e6678c5f-ceph\") pod \"3a4ab9f9-573b-4953-8683-1940e6678c5f\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.760523 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-internal-tls-certs\") pod \"3a4ab9f9-573b-4953-8683-1940e6678c5f\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.760805 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a4ab9f9-573b-4953-8683-1940e6678c5f-logs\") pod \"3a4ab9f9-573b-4953-8683-1940e6678c5f\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.760837 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvjcf\" (UniqueName: \"kubernetes.io/projected/3a4ab9f9-573b-4953-8683-1940e6678c5f-kube-api-access-xvjcf\") pod \"3a4ab9f9-573b-4953-8683-1940e6678c5f\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.762003 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a4ab9f9-573b-4953-8683-1940e6678c5f-logs" (OuterVolumeSpecName: "logs") pod "3a4ab9f9-573b-4953-8683-1940e6678c5f" (UID: "3a4ab9f9-573b-4953-8683-1940e6678c5f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.763813 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.766723 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a4ab9f9-573b-4953-8683-1940e6678c5f-kube-api-access-xvjcf" (OuterVolumeSpecName: "kube-api-access-xvjcf") pod "3a4ab9f9-573b-4953-8683-1940e6678c5f" (UID: "3a4ab9f9-573b-4953-8683-1940e6678c5f"). InnerVolumeSpecName "kube-api-access-xvjcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.767133 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a4ab9f9-573b-4953-8683-1940e6678c5f-ceph" (OuterVolumeSpecName: "ceph") pod "3a4ab9f9-573b-4953-8683-1940e6678c5f" (UID: "3a4ab9f9-573b-4953-8683-1940e6678c5f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.769077 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "3a4ab9f9-573b-4953-8683-1940e6678c5f" (UID: "3a4ab9f9-573b-4953-8683-1940e6678c5f"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.818233 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3a4ab9f9-573b-4953-8683-1940e6678c5f" (UID: "3a4ab9f9-573b-4953-8683-1940e6678c5f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.862962 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf03a664-65df-4df2-a074-d35397ff3913-logs\") pod \"bf03a664-65df-4df2-a074-d35397ff3913\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.863227 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/bf03a664-65df-4df2-a074-d35397ff3913-ceph\") pod \"bf03a664-65df-4df2-a074-d35397ff3913\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.863314 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-combined-ca-bundle\") pod \"bf03a664-65df-4df2-a074-d35397ff3913\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.863389 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-config-data\") pod \"bf03a664-65df-4df2-a074-d35397ff3913\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.863406 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf03a664-65df-4df2-a074-d35397ff3913-httpd-run\") pod \"bf03a664-65df-4df2-a074-d35397ff3913\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.863448 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-scripts\") pod \"bf03a664-65df-4df2-a074-d35397ff3913\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.863472 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3a4ab9f9-573b-4953-8683-1940e6678c5f-httpd-run\") pod \"3a4ab9f9-573b-4953-8683-1940e6678c5f\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.863538 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-scripts\") pod \"3a4ab9f9-573b-4953-8683-1940e6678c5f\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.863574 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-public-tls-certs\") pod \"bf03a664-65df-4df2-a074-d35397ff3913\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.863600 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-config-data\") pod \"3a4ab9f9-573b-4953-8683-1940e6678c5f\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.863639 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"bf03a664-65df-4df2-a074-d35397ff3913\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.863642 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf03a664-65df-4df2-a074-d35397ff3913-logs" (OuterVolumeSpecName: "logs") pod "bf03a664-65df-4df2-a074-d35397ff3913" (UID: "bf03a664-65df-4df2-a074-d35397ff3913"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.863656 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wn2wh\" (UniqueName: \"kubernetes.io/projected/bf03a664-65df-4df2-a074-d35397ff3913-kube-api-access-wn2wh\") pod \"bf03a664-65df-4df2-a074-d35397ff3913\" (UID: \"bf03a664-65df-4df2-a074-d35397ff3913\") " Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.863736 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-combined-ca-bundle\") pod \"3a4ab9f9-573b-4953-8683-1940e6678c5f\" (UID: \"3a4ab9f9-573b-4953-8683-1940e6678c5f\") " Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.863846 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf03a664-65df-4df2-a074-d35397ff3913-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bf03a664-65df-4df2-a074-d35397ff3913" (UID: "bf03a664-65df-4df2-a074-d35397ff3913"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.864593 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a4ab9f9-573b-4953-8683-1940e6678c5f-logs\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.864609 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvjcf\" (UniqueName: \"kubernetes.io/projected/3a4ab9f9-573b-4953-8683-1940e6678c5f-kube-api-access-xvjcf\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.864619 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf03a664-65df-4df2-a074-d35397ff3913-logs\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.864627 4818 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf03a664-65df-4df2-a074-d35397ff3913-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.864647 4818 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.864655 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3a4ab9f9-573b-4953-8683-1940e6678c5f-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.864664 4818 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.864657 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a4ab9f9-573b-4953-8683-1940e6678c5f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3a4ab9f9-573b-4953-8683-1940e6678c5f" (UID: "3a4ab9f9-573b-4953-8683-1940e6678c5f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.866777 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-scripts" (OuterVolumeSpecName: "scripts") pod "bf03a664-65df-4df2-a074-d35397ff3913" (UID: "bf03a664-65df-4df2-a074-d35397ff3913"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.867926 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "bf03a664-65df-4df2-a074-d35397ff3913" (UID: "bf03a664-65df-4df2-a074-d35397ff3913"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.870872 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-scripts" (OuterVolumeSpecName: "scripts") pod "3a4ab9f9-573b-4953-8683-1940e6678c5f" (UID: "3a4ab9f9-573b-4953-8683-1940e6678c5f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.871391 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf03a664-65df-4df2-a074-d35397ff3913-kube-api-access-wn2wh" (OuterVolumeSpecName: "kube-api-access-wn2wh") pod "bf03a664-65df-4df2-a074-d35397ff3913" (UID: "bf03a664-65df-4df2-a074-d35397ff3913"). InnerVolumeSpecName "kube-api-access-wn2wh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.871722 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf03a664-65df-4df2-a074-d35397ff3913-ceph" (OuterVolumeSpecName: "ceph") pod "bf03a664-65df-4df2-a074-d35397ff3913" (UID: "bf03a664-65df-4df2-a074-d35397ff3913"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.895350 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf03a664-65df-4df2-a074-d35397ff3913" (UID: "bf03a664-65df-4df2-a074-d35397ff3913"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.895382 4818 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.903430 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a4ab9f9-573b-4953-8683-1940e6678c5f" (UID: "3a4ab9f9-573b-4953-8683-1940e6678c5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.921534 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-config-data" (OuterVolumeSpecName: "config-data") pod "bf03a664-65df-4df2-a074-d35397ff3913" (UID: "bf03a664-65df-4df2-a074-d35397ff3913"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.927461 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "bf03a664-65df-4df2-a074-d35397ff3913" (UID: "bf03a664-65df-4df2-a074-d35397ff3913"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.928767 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-config-data" (OuterVolumeSpecName: "config-data") pod "3a4ab9f9-573b-4953-8683-1940e6678c5f" (UID: "3a4ab9f9-573b-4953-8683-1940e6678c5f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.966747 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/bf03a664-65df-4df2-a074-d35397ff3913-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.966779 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.966789 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.966797 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.966811 4818 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3a4ab9f9-573b-4953-8683-1940e6678c5f-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.966820 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.966828 4818 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf03a664-65df-4df2-a074-d35397ff3913-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.966836 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.966865 4818 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.966876 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wn2wh\" (UniqueName: \"kubernetes.io/projected/bf03a664-65df-4df2-a074-d35397ff3913-kube-api-access-wn2wh\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.966885 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a4ab9f9-573b-4953-8683-1940e6678c5f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.966893 4818 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:34 crc kubenswrapper[4818]: I1122 05:45:34.984834 4818 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.021864 4818 generic.go:334] "Generic (PLEG): container finished" podID="3a4ab9f9-573b-4953-8683-1940e6678c5f" containerID="9298b0c1209abe09e04105d98507048a99335a51312bda567ab76751b8da4c9c" exitCode=0 Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.021899 4818 generic.go:334] "Generic (PLEG): container finished" podID="3a4ab9f9-573b-4953-8683-1940e6678c5f" containerID="f90bd608574b9d6050a979456443e31d4dcd782f7b71be215290a26a30506b74" exitCode=143 Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.021923 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.021996 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3a4ab9f9-573b-4953-8683-1940e6678c5f","Type":"ContainerDied","Data":"9298b0c1209abe09e04105d98507048a99335a51312bda567ab76751b8da4c9c"} Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.022027 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3a4ab9f9-573b-4953-8683-1940e6678c5f","Type":"ContainerDied","Data":"f90bd608574b9d6050a979456443e31d4dcd782f7b71be215290a26a30506b74"} Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.022037 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3a4ab9f9-573b-4953-8683-1940e6678c5f","Type":"ContainerDied","Data":"267ac408cba6aff5449637b3c3e47043f9452c11e2d06dff1f51b5dd093b4b2f"} Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.022054 4818 scope.go:117] "RemoveContainer" containerID="9298b0c1209abe09e04105d98507048a99335a51312bda567ab76751b8da4c9c" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.024874 4818 generic.go:334] "Generic (PLEG): container finished" podID="bf03a664-65df-4df2-a074-d35397ff3913" containerID="e0891fe59ab42617392296a9421c8e855213d8a70c68190080712cd69e5b56ce" exitCode=0 Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.024906 4818 generic.go:334] "Generic (PLEG): container finished" podID="bf03a664-65df-4df2-a074-d35397ff3913" containerID="78683020f40cca0195bb98a680aa641ab7b2b9ed9a3b815213f6f1c7af2909fc" exitCode=143 Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.024980 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bf03a664-65df-4df2-a074-d35397ff3913","Type":"ContainerDied","Data":"e0891fe59ab42617392296a9421c8e855213d8a70c68190080712cd69e5b56ce"} Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.025009 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bf03a664-65df-4df2-a074-d35397ff3913","Type":"ContainerDied","Data":"78683020f40cca0195bb98a680aa641ab7b2b9ed9a3b815213f6f1c7af2909fc"} Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.025021 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bf03a664-65df-4df2-a074-d35397ff3913","Type":"ContainerDied","Data":"788cede5865c6c22804ff1125f51dff431a47b52687dd24668a1b35d6affde47"} Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.024898 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.026518 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-2xttz" event={"ID":"64f4e98d-2953-472c-852a-424b72c513a9","Type":"ContainerStarted","Data":"13bb3ccd64b4270bb8706864751f2ab55a499d5559887785e193a1dc1f688b57"} Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.060074 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.069235 4818 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.073819 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.132586 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 05:45:35 crc kubenswrapper[4818]: E1122 05:45:35.133012 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a4ab9f9-573b-4953-8683-1940e6678c5f" containerName="glance-log" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.133027 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a4ab9f9-573b-4953-8683-1940e6678c5f" containerName="glance-log" Nov 22 05:45:35 crc kubenswrapper[4818]: E1122 05:45:35.133041 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf03a664-65df-4df2-a074-d35397ff3913" containerName="glance-log" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.133055 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf03a664-65df-4df2-a074-d35397ff3913" containerName="glance-log" Nov 22 05:45:35 crc kubenswrapper[4818]: E1122 05:45:35.133073 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf03a664-65df-4df2-a074-d35397ff3913" containerName="glance-httpd" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.133079 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf03a664-65df-4df2-a074-d35397ff3913" containerName="glance-httpd" Nov 22 05:45:35 crc kubenswrapper[4818]: E1122 05:45:35.133110 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a4ab9f9-573b-4953-8683-1940e6678c5f" containerName="glance-httpd" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.133116 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a4ab9f9-573b-4953-8683-1940e6678c5f" containerName="glance-httpd" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.133346 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf03a664-65df-4df2-a074-d35397ff3913" containerName="glance-httpd" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.133365 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a4ab9f9-573b-4953-8683-1940e6678c5f" containerName="glance-httpd" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.133375 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a4ab9f9-573b-4953-8683-1940e6678c5f" containerName="glance-log" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.133391 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf03a664-65df-4df2-a074-d35397ff3913" containerName="glance-log" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.134410 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.137590 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.137907 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.137986 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.138069 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2zbds" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.141998 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.163619 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.176718 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.186604 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.188649 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.190544 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.190843 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.200081 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.274602 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f87508a5-9796-46dd-b9a4-29815fd608ad-logs\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.274645 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f87508a5-9796-46dd-b9a4-29815fd608ad-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.274687 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f87508a5-9796-46dd-b9a4-29815fd608ad-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.274800 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f87508a5-9796-46dd-b9a4-29815fd608ad-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.274859 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f87508a5-9796-46dd-b9a4-29815fd608ad-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.274919 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxd5t\" (UniqueName: \"kubernetes.io/projected/f87508a5-9796-46dd-b9a4-29815fd608ad-kube-api-access-hxd5t\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.275040 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f87508a5-9796-46dd-b9a4-29815fd608ad-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.275099 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f87508a5-9796-46dd-b9a4-29815fd608ad-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.275137 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.377422 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6652494e-6fe5-4c2b-b265-bb2e5567265e-logs\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.377504 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f87508a5-9796-46dd-b9a4-29815fd608ad-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.377628 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.377660 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f87508a5-9796-46dd-b9a4-29815fd608ad-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.377680 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6652494e-6fe5-4c2b-b265-bb2e5567265e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.377728 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.377768 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6652494e-6fe5-4c2b-b265-bb2e5567265e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.377817 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f87508a5-9796-46dd-b9a4-29815fd608ad-logs\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.377840 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f87508a5-9796-46dd-b9a4-29815fd608ad-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.377863 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6652494e-6fe5-4c2b-b265-bb2e5567265e-scripts\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.377912 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.377915 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f87508a5-9796-46dd-b9a4-29815fd608ad-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.377995 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6652494e-6fe5-4c2b-b265-bb2e5567265e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.378025 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f87508a5-9796-46dd-b9a4-29815fd608ad-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.378053 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f87508a5-9796-46dd-b9a4-29815fd608ad-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.378078 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6652494e-6fe5-4c2b-b265-bb2e5567265e-ceph\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.378125 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxd5t\" (UniqueName: \"kubernetes.io/projected/f87508a5-9796-46dd-b9a4-29815fd608ad-kube-api-access-hxd5t\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.378152 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f87508a5-9796-46dd-b9a4-29815fd608ad-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.378157 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwzff\" (UniqueName: \"kubernetes.io/projected/6652494e-6fe5-4c2b-b265-bb2e5567265e-kube-api-access-wwzff\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.378240 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6652494e-6fe5-4c2b-b265-bb2e5567265e-config-data\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.378884 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f87508a5-9796-46dd-b9a4-29815fd608ad-logs\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.382041 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f87508a5-9796-46dd-b9a4-29815fd608ad-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.383135 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f87508a5-9796-46dd-b9a4-29815fd608ad-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.383509 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f87508a5-9796-46dd-b9a4-29815fd608ad-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.387045 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f87508a5-9796-46dd-b9a4-29815fd608ad-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.388860 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f87508a5-9796-46dd-b9a4-29815fd608ad-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.394133 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxd5t\" (UniqueName: \"kubernetes.io/projected/f87508a5-9796-46dd-b9a4-29815fd608ad-kube-api-access-hxd5t\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.405916 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"f87508a5-9796-46dd-b9a4-29815fd608ad\") " pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.461880 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.480421 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.480469 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6652494e-6fe5-4c2b-b265-bb2e5567265e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.480573 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6652494e-6fe5-4c2b-b265-bb2e5567265e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.480614 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.480626 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6652494e-6fe5-4c2b-b265-bb2e5567265e-scripts\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.481586 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6652494e-6fe5-4c2b-b265-bb2e5567265e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.481617 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6652494e-6fe5-4c2b-b265-bb2e5567265e-ceph\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.481687 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwzff\" (UniqueName: \"kubernetes.io/projected/6652494e-6fe5-4c2b-b265-bb2e5567265e-kube-api-access-wwzff\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.481775 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6652494e-6fe5-4c2b-b265-bb2e5567265e-config-data\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.481853 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6652494e-6fe5-4c2b-b265-bb2e5567265e-logs\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.482278 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6652494e-6fe5-4c2b-b265-bb2e5567265e-logs\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.483784 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6652494e-6fe5-4c2b-b265-bb2e5567265e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.485323 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6652494e-6fe5-4c2b-b265-bb2e5567265e-scripts\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.487446 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6652494e-6fe5-4c2b-b265-bb2e5567265e-config-data\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.488059 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6652494e-6fe5-4c2b-b265-bb2e5567265e-ceph\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.490837 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6652494e-6fe5-4c2b-b265-bb2e5567265e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.494411 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6652494e-6fe5-4c2b-b265-bb2e5567265e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.500423 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwzff\" (UniqueName: \"kubernetes.io/projected/6652494e-6fe5-4c2b-b265-bb2e5567265e-kube-api-access-wwzff\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.545544 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"6652494e-6fe5-4c2b-b265-bb2e5567265e\") " pod="openstack/glance-default-external-api-0" Nov 22 05:45:35 crc kubenswrapper[4818]: I1122 05:45:35.810120 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 05:45:36 crc kubenswrapper[4818]: I1122 05:45:36.303888 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a4ab9f9-573b-4953-8683-1940e6678c5f" path="/var/lib/kubelet/pods/3a4ab9f9-573b-4953-8683-1940e6678c5f/volumes" Nov 22 05:45:36 crc kubenswrapper[4818]: I1122 05:45:36.304854 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf03a664-65df-4df2-a074-d35397ff3913" path="/var/lib/kubelet/pods/bf03a664-65df-4df2-a074-d35397ff3913/volumes" Nov 22 05:45:37 crc kubenswrapper[4818]: I1122 05:45:37.546571 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ksfkm" Nov 22 05:45:37 crc kubenswrapper[4818]: I1122 05:45:37.547031 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ksfkm" Nov 22 05:45:37 crc kubenswrapper[4818]: I1122 05:45:37.722523 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Nov 22 05:45:37 crc kubenswrapper[4818]: I1122 05:45:37.965527 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Nov 22 05:45:38 crc kubenswrapper[4818]: I1122 05:45:38.649831 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-ksfkm" podUID="e3759d20-d679-42de-9a4e-f32305ec1ca0" containerName="registry-server" probeResult="failure" output=< Nov 22 05:45:38 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Nov 22 05:45:38 crc kubenswrapper[4818]: > Nov 22 05:45:41 crc kubenswrapper[4818]: I1122 05:45:41.097383 4818 scope.go:117] "RemoveContainer" containerID="2ff61c4f9d6d40fc8a93a3eedf71ee13fa18a916b73695691c0b347cc8abee1a" Nov 22 05:45:42 crc kubenswrapper[4818]: E1122 05:45:42.024905 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3759d20_d679_42de_9a4e_f32305ec1ca0.slice/crio-conmon-aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3759d20_d679_42de_9a4e_f32305ec1ca0.slice/crio-aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514.scope\": RecentStats: unable to find data in memory cache]" Nov 22 05:45:43 crc kubenswrapper[4818]: I1122 05:45:43.169091 4818 scope.go:117] "RemoveContainer" containerID="f90bd608574b9d6050a979456443e31d4dcd782f7b71be215290a26a30506b74" Nov 22 05:45:43 crc kubenswrapper[4818]: I1122 05:45:43.287027 4818 scope.go:117] "RemoveContainer" containerID="9298b0c1209abe09e04105d98507048a99335a51312bda567ab76751b8da4c9c" Nov 22 05:45:43 crc kubenswrapper[4818]: E1122 05:45:43.289520 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9298b0c1209abe09e04105d98507048a99335a51312bda567ab76751b8da4c9c\": container with ID starting with 9298b0c1209abe09e04105d98507048a99335a51312bda567ab76751b8da4c9c not found: ID does not exist" containerID="9298b0c1209abe09e04105d98507048a99335a51312bda567ab76751b8da4c9c" Nov 22 05:45:43 crc kubenswrapper[4818]: I1122 05:45:43.289558 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9298b0c1209abe09e04105d98507048a99335a51312bda567ab76751b8da4c9c"} err="failed to get container status \"9298b0c1209abe09e04105d98507048a99335a51312bda567ab76751b8da4c9c\": rpc error: code = NotFound desc = could not find container \"9298b0c1209abe09e04105d98507048a99335a51312bda567ab76751b8da4c9c\": container with ID starting with 9298b0c1209abe09e04105d98507048a99335a51312bda567ab76751b8da4c9c not found: ID does not exist" Nov 22 05:45:43 crc kubenswrapper[4818]: I1122 05:45:43.289583 4818 scope.go:117] "RemoveContainer" containerID="f90bd608574b9d6050a979456443e31d4dcd782f7b71be215290a26a30506b74" Nov 22 05:45:43 crc kubenswrapper[4818]: E1122 05:45:43.290154 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f90bd608574b9d6050a979456443e31d4dcd782f7b71be215290a26a30506b74\": container with ID starting with f90bd608574b9d6050a979456443e31d4dcd782f7b71be215290a26a30506b74 not found: ID does not exist" containerID="f90bd608574b9d6050a979456443e31d4dcd782f7b71be215290a26a30506b74" Nov 22 05:45:43 crc kubenswrapper[4818]: I1122 05:45:43.290194 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f90bd608574b9d6050a979456443e31d4dcd782f7b71be215290a26a30506b74"} err="failed to get container status \"f90bd608574b9d6050a979456443e31d4dcd782f7b71be215290a26a30506b74\": rpc error: code = NotFound desc = could not find container \"f90bd608574b9d6050a979456443e31d4dcd782f7b71be215290a26a30506b74\": container with ID starting with f90bd608574b9d6050a979456443e31d4dcd782f7b71be215290a26a30506b74 not found: ID does not exist" Nov 22 05:45:43 crc kubenswrapper[4818]: I1122 05:45:43.290277 4818 scope.go:117] "RemoveContainer" containerID="9298b0c1209abe09e04105d98507048a99335a51312bda567ab76751b8da4c9c" Nov 22 05:45:43 crc kubenswrapper[4818]: I1122 05:45:43.290737 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9298b0c1209abe09e04105d98507048a99335a51312bda567ab76751b8da4c9c"} err="failed to get container status \"9298b0c1209abe09e04105d98507048a99335a51312bda567ab76751b8da4c9c\": rpc error: code = NotFound desc = could not find container \"9298b0c1209abe09e04105d98507048a99335a51312bda567ab76751b8da4c9c\": container with ID starting with 9298b0c1209abe09e04105d98507048a99335a51312bda567ab76751b8da4c9c not found: ID does not exist" Nov 22 05:45:43 crc kubenswrapper[4818]: I1122 05:45:43.290765 4818 scope.go:117] "RemoveContainer" containerID="f90bd608574b9d6050a979456443e31d4dcd782f7b71be215290a26a30506b74" Nov 22 05:45:43 crc kubenswrapper[4818]: I1122 05:45:43.291213 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f90bd608574b9d6050a979456443e31d4dcd782f7b71be215290a26a30506b74"} err="failed to get container status \"f90bd608574b9d6050a979456443e31d4dcd782f7b71be215290a26a30506b74\": rpc error: code = NotFound desc = could not find container \"f90bd608574b9d6050a979456443e31d4dcd782f7b71be215290a26a30506b74\": container with ID starting with f90bd608574b9d6050a979456443e31d4dcd782f7b71be215290a26a30506b74 not found: ID does not exist" Nov 22 05:45:43 crc kubenswrapper[4818]: I1122 05:45:43.291230 4818 scope.go:117] "RemoveContainer" containerID="e0891fe59ab42617392296a9421c8e855213d8a70c68190080712cd69e5b56ce" Nov 22 05:45:43 crc kubenswrapper[4818]: I1122 05:45:43.514585 4818 scope.go:117] "RemoveContainer" containerID="78683020f40cca0195bb98a680aa641ab7b2b9ed9a3b815213f6f1c7af2909fc" Nov 22 05:45:43 crc kubenswrapper[4818]: I1122 05:45:43.543837 4818 scope.go:117] "RemoveContainer" containerID="e0891fe59ab42617392296a9421c8e855213d8a70c68190080712cd69e5b56ce" Nov 22 05:45:43 crc kubenswrapper[4818]: E1122 05:45:43.544445 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0891fe59ab42617392296a9421c8e855213d8a70c68190080712cd69e5b56ce\": container with ID starting with e0891fe59ab42617392296a9421c8e855213d8a70c68190080712cd69e5b56ce not found: ID does not exist" containerID="e0891fe59ab42617392296a9421c8e855213d8a70c68190080712cd69e5b56ce" Nov 22 05:45:43 crc kubenswrapper[4818]: I1122 05:45:43.544493 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0891fe59ab42617392296a9421c8e855213d8a70c68190080712cd69e5b56ce"} err="failed to get container status \"e0891fe59ab42617392296a9421c8e855213d8a70c68190080712cd69e5b56ce\": rpc error: code = NotFound desc = could not find container \"e0891fe59ab42617392296a9421c8e855213d8a70c68190080712cd69e5b56ce\": container with ID starting with e0891fe59ab42617392296a9421c8e855213d8a70c68190080712cd69e5b56ce not found: ID does not exist" Nov 22 05:45:43 crc kubenswrapper[4818]: I1122 05:45:43.544521 4818 scope.go:117] "RemoveContainer" containerID="78683020f40cca0195bb98a680aa641ab7b2b9ed9a3b815213f6f1c7af2909fc" Nov 22 05:45:43 crc kubenswrapper[4818]: E1122 05:45:43.544916 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78683020f40cca0195bb98a680aa641ab7b2b9ed9a3b815213f6f1c7af2909fc\": container with ID starting with 78683020f40cca0195bb98a680aa641ab7b2b9ed9a3b815213f6f1c7af2909fc not found: ID does not exist" containerID="78683020f40cca0195bb98a680aa641ab7b2b9ed9a3b815213f6f1c7af2909fc" Nov 22 05:45:43 crc kubenswrapper[4818]: I1122 05:45:43.544965 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78683020f40cca0195bb98a680aa641ab7b2b9ed9a3b815213f6f1c7af2909fc"} err="failed to get container status \"78683020f40cca0195bb98a680aa641ab7b2b9ed9a3b815213f6f1c7af2909fc\": rpc error: code = NotFound desc = could not find container \"78683020f40cca0195bb98a680aa641ab7b2b9ed9a3b815213f6f1c7af2909fc\": container with ID starting with 78683020f40cca0195bb98a680aa641ab7b2b9ed9a3b815213f6f1c7af2909fc not found: ID does not exist" Nov 22 05:45:43 crc kubenswrapper[4818]: I1122 05:45:43.544994 4818 scope.go:117] "RemoveContainer" containerID="e0891fe59ab42617392296a9421c8e855213d8a70c68190080712cd69e5b56ce" Nov 22 05:45:43 crc kubenswrapper[4818]: I1122 05:45:43.545318 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0891fe59ab42617392296a9421c8e855213d8a70c68190080712cd69e5b56ce"} err="failed to get container status \"e0891fe59ab42617392296a9421c8e855213d8a70c68190080712cd69e5b56ce\": rpc error: code = NotFound desc = could not find container \"e0891fe59ab42617392296a9421c8e855213d8a70c68190080712cd69e5b56ce\": container with ID starting with e0891fe59ab42617392296a9421c8e855213d8a70c68190080712cd69e5b56ce not found: ID does not exist" Nov 22 05:45:43 crc kubenswrapper[4818]: I1122 05:45:43.545350 4818 scope.go:117] "RemoveContainer" containerID="78683020f40cca0195bb98a680aa641ab7b2b9ed9a3b815213f6f1c7af2909fc" Nov 22 05:45:43 crc kubenswrapper[4818]: I1122 05:45:43.545573 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78683020f40cca0195bb98a680aa641ab7b2b9ed9a3b815213f6f1c7af2909fc"} err="failed to get container status \"78683020f40cca0195bb98a680aa641ab7b2b9ed9a3b815213f6f1c7af2909fc\": rpc error: code = NotFound desc = could not find container \"78683020f40cca0195bb98a680aa641ab7b2b9ed9a3b815213f6f1c7af2909fc\": container with ID starting with 78683020f40cca0195bb98a680aa641ab7b2b9ed9a3b815213f6f1c7af2909fc not found: ID does not exist" Nov 22 05:45:43 crc kubenswrapper[4818]: I1122 05:45:43.923740 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 05:45:43 crc kubenswrapper[4818]: W1122 05:45:43.940467 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6652494e_6fe5_4c2b_b265_bb2e5567265e.slice/crio-95a1e1535749a33d4ef4fda9e735639fe8e64e78af161a2976f796e772f2f6c4 WatchSource:0}: Error finding container 95a1e1535749a33d4ef4fda9e735639fe8e64e78af161a2976f796e772f2f6c4: Status 404 returned error can't find the container with id 95a1e1535749a33d4ef4fda9e735639fe8e64e78af161a2976f796e772f2f6c4 Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.025648 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 05:45:44 crc kubenswrapper[4818]: W1122 05:45:44.037675 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf87508a5_9796_46dd_b9a4_29815fd608ad.slice/crio-1750cafba9980f47040d5a7509c84f8dfc736f3571502b2c0e7c2846f9d3c053 WatchSource:0}: Error finding container 1750cafba9980f47040d5a7509c84f8dfc736f3571502b2c0e7c2846f9d3c053: Status 404 returned error can't find the container with id 1750cafba9980f47040d5a7509c84f8dfc736f3571502b2c0e7c2846f9d3c053 Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.149187 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6652494e-6fe5-4c2b-b265-bb2e5567265e","Type":"ContainerStarted","Data":"95a1e1535749a33d4ef4fda9e735639fe8e64e78af161a2976f796e772f2f6c4"} Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.152045 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f784b6cfc-t55xf" event={"ID":"d03bb0b9-32cd-43b8-92e4-2b8f83361246","Type":"ContainerStarted","Data":"fb8ef4485cc8560984370390d18efaff10c06a33c0b320e64cb2d496a0f8109b"} Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.152109 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f784b6cfc-t55xf" event={"ID":"d03bb0b9-32cd-43b8-92e4-2b8f83361246","Type":"ContainerStarted","Data":"03e2148476f595fa4c74ff9412cdb8bfb41d861a81114fa9b7642fafe4b21c70"} Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.152136 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-f784b6cfc-t55xf" podUID="d03bb0b9-32cd-43b8-92e4-2b8f83361246" containerName="horizon-log" containerID="cri-o://03e2148476f595fa4c74ff9412cdb8bfb41d861a81114fa9b7642fafe4b21c70" gracePeriod=30 Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.152218 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-f784b6cfc-t55xf" podUID="d03bb0b9-32cd-43b8-92e4-2b8f83361246" containerName="horizon" containerID="cri-o://fb8ef4485cc8560984370390d18efaff10c06a33c0b320e64cb2d496a0f8109b" gracePeriod=30 Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.176859 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86757bfcb-s9bbk" event={"ID":"34f2acb8-906b-4520-86c2-bb603a964489","Type":"ContainerStarted","Data":"95cf46d995ac196336458c077405d03f9e5a0be93f89c30ae16977376a676314"} Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.176901 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86757bfcb-s9bbk" event={"ID":"34f2acb8-906b-4520-86c2-bb603a964489","Type":"ContainerStarted","Data":"b8140c3c5ffa3533051f6e7927b30d6f697de2d555b242bb4eb6e0d9bcc1e47d"} Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.181619 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-2xttz" event={"ID":"64f4e98d-2953-472c-852a-424b72c513a9","Type":"ContainerStarted","Data":"efd6a02edfb01d2064fc0e366c29accec0a100d9d2b56e5807f6c968832182b8"} Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.184709 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f87508a5-9796-46dd-b9a4-29815fd608ad","Type":"ContainerStarted","Data":"1750cafba9980f47040d5a7509c84f8dfc736f3571502b2c0e7c2846f9d3c053"} Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.188437 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c4458788c-rwwjv" event={"ID":"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8","Type":"ContainerStarted","Data":"39f6a96f45c4ddd4908190f409436048cbd94f00d707a0d23d84995b34baa4ef"} Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.188475 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c4458788c-rwwjv" event={"ID":"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8","Type":"ContainerStarted","Data":"c169439386f8d7a558f2d85a41494b3e0548e5681b0841253f085f632029c464"} Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.188602 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-c4458788c-rwwjv" podUID="90e9c626-fc6c-4ad9-b385-e634bd3dbfb8" containerName="horizon-log" containerID="cri-o://c169439386f8d7a558f2d85a41494b3e0548e5681b0841253f085f632029c464" gracePeriod=30 Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.188630 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-c4458788c-rwwjv" podUID="90e9c626-fc6c-4ad9-b385-e634bd3dbfb8" containerName="horizon" containerID="cri-o://39f6a96f45c4ddd4908190f409436048cbd94f00d707a0d23d84995b34baa4ef" gracePeriod=30 Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.194929 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7764859bb4-tnzkk" event={"ID":"60b9da36-1281-4cc5-a9ac-a2df2e4d9638","Type":"ContainerStarted","Data":"64ff53e988010bba9bf54bb360558aabbc0e90386763b14c6ab5fe8f007de1e8"} Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.195619 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7764859bb4-tnzkk" event={"ID":"60b9da36-1281-4cc5-a9ac-a2df2e4d9638","Type":"ContainerStarted","Data":"180e7581aa51496380128694e8a8eda047b65550ed4a7f9e238af15906922636"} Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.206814 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-f784b6cfc-t55xf" podStartSLOduration=2.895294705 podStartE2EDuration="17.206796552s" podCreationTimestamp="2025-11-22 05:45:27 +0000 UTC" firstStartedPulling="2025-11-22 05:45:28.975813744 +0000 UTC m=+3481.550230271" lastFinishedPulling="2025-11-22 05:45:43.287315591 +0000 UTC m=+3495.861732118" observedRunningTime="2025-11-22 05:45:44.170045453 +0000 UTC m=+3496.744461980" watchObservedRunningTime="2025-11-22 05:45:44.206796552 +0000 UTC m=+3496.781213079" Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.208863 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-86757bfcb-s9bbk" podStartSLOduration=2.964636825 podStartE2EDuration="14.208851986s" podCreationTimestamp="2025-11-22 05:45:30 +0000 UTC" firstStartedPulling="2025-11-22 05:45:32.043350426 +0000 UTC m=+3484.617766953" lastFinishedPulling="2025-11-22 05:45:43.287565587 +0000 UTC m=+3495.861982114" observedRunningTime="2025-11-22 05:45:44.197893745 +0000 UTC m=+3496.772310272" watchObservedRunningTime="2025-11-22 05:45:44.208851986 +0000 UTC m=+3496.783268513" Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.222128 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-2xttz" podStartSLOduration=1.8775234379999999 podStartE2EDuration="11.22211123s" podCreationTimestamp="2025-11-22 05:45:33 +0000 UTC" firstStartedPulling="2025-11-22 05:45:34.000708915 +0000 UTC m=+3486.575125452" lastFinishedPulling="2025-11-22 05:45:43.345296707 +0000 UTC m=+3495.919713244" observedRunningTime="2025-11-22 05:45:44.213752368 +0000 UTC m=+3496.788168915" watchObservedRunningTime="2025-11-22 05:45:44.22211123 +0000 UTC m=+3496.796527757" Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.240974 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-c4458788c-rwwjv" podStartSLOduration=2.02118581 podStartE2EDuration="16.240953502s" podCreationTimestamp="2025-11-22 05:45:28 +0000 UTC" firstStartedPulling="2025-11-22 05:45:29.12683203 +0000 UTC m=+3481.701248557" lastFinishedPulling="2025-11-22 05:45:43.346599722 +0000 UTC m=+3495.921016249" observedRunningTime="2025-11-22 05:45:44.232395564 +0000 UTC m=+3496.806812091" watchObservedRunningTime="2025-11-22 05:45:44.240953502 +0000 UTC m=+3496.815370029" Nov 22 05:45:44 crc kubenswrapper[4818]: I1122 05:45:44.256089 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7764859bb4-tnzkk" podStartSLOduration=2.64372426 podStartE2EDuration="14.256066605s" podCreationTimestamp="2025-11-22 05:45:30 +0000 UTC" firstStartedPulling="2025-11-22 05:45:31.731302748 +0000 UTC m=+3484.305719275" lastFinishedPulling="2025-11-22 05:45:43.343645093 +0000 UTC m=+3495.918061620" observedRunningTime="2025-11-22 05:45:44.248172365 +0000 UTC m=+3496.822588902" watchObservedRunningTime="2025-11-22 05:45:44.256066605 +0000 UTC m=+3496.830483132" Nov 22 05:45:45 crc kubenswrapper[4818]: I1122 05:45:45.223969 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f87508a5-9796-46dd-b9a4-29815fd608ad","Type":"ContainerStarted","Data":"e3ee9fd54897ca7909ea22ee3a3b897bff514251c4e48977a3cf6176f942ea3a"} Nov 22 05:45:45 crc kubenswrapper[4818]: I1122 05:45:45.227545 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6652494e-6fe5-4c2b-b265-bb2e5567265e","Type":"ContainerStarted","Data":"2dda94da79ae1635874a6b34cf7c807c86c04b88d03bf44c85052888ead8d4aa"} Nov 22 05:45:45 crc kubenswrapper[4818]: I1122 05:45:45.227603 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6652494e-6fe5-4c2b-b265-bb2e5567265e","Type":"ContainerStarted","Data":"7ab793211994584be16453462dd9dc0cb26ec59f04d00d9e59601dc530d6c3b6"} Nov 22 05:45:45 crc kubenswrapper[4818]: I1122 05:45:45.261183 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=10.261164469 podStartE2EDuration="10.261164469s" podCreationTimestamp="2025-11-22 05:45:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:45:45.251116071 +0000 UTC m=+3497.825532618" watchObservedRunningTime="2025-11-22 05:45:45.261164469 +0000 UTC m=+3497.835580996" Nov 22 05:45:45 crc kubenswrapper[4818]: I1122 05:45:45.811101 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 22 05:45:45 crc kubenswrapper[4818]: I1122 05:45:45.811155 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 22 05:45:45 crc kubenswrapper[4818]: I1122 05:45:45.878018 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 22 05:45:45 crc kubenswrapper[4818]: I1122 05:45:45.887923 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 22 05:45:46 crc kubenswrapper[4818]: I1122 05:45:46.242604 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f87508a5-9796-46dd-b9a4-29815fd608ad","Type":"ContainerStarted","Data":"66464f74b35bda4719788b36ce0ec37a7e8b5bb3ad1fd718d48c6ff6198f9ef1"} Nov 22 05:45:46 crc kubenswrapper[4818]: I1122 05:45:46.243102 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 22 05:45:46 crc kubenswrapper[4818]: I1122 05:45:46.243130 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 22 05:45:46 crc kubenswrapper[4818]: I1122 05:45:46.265140 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=11.265122252 podStartE2EDuration="11.265122252s" podCreationTimestamp="2025-11-22 05:45:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:45:46.264559507 +0000 UTC m=+3498.838976054" watchObservedRunningTime="2025-11-22 05:45:46.265122252 +0000 UTC m=+3498.839538779" Nov 22 05:45:48 crc kubenswrapper[4818]: I1122 05:45:48.417853 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-f784b6cfc-t55xf" Nov 22 05:45:48 crc kubenswrapper[4818]: I1122 05:45:48.589048 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-c4458788c-rwwjv" Nov 22 05:45:48 crc kubenswrapper[4818]: I1122 05:45:48.612558 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-ksfkm" podUID="e3759d20-d679-42de-9a4e-f32305ec1ca0" containerName="registry-server" probeResult="failure" output=< Nov 22 05:45:48 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Nov 22 05:45:48 crc kubenswrapper[4818]: > Nov 22 05:45:51 crc kubenswrapper[4818]: I1122 05:45:51.101835 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:51 crc kubenswrapper[4818]: I1122 05:45:51.102815 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:45:51 crc kubenswrapper[4818]: I1122 05:45:51.168052 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-655mn"] Nov 22 05:45:51 crc kubenswrapper[4818]: I1122 05:45:51.172909 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-655mn" Nov 22 05:45:51 crc kubenswrapper[4818]: I1122 05:45:51.188711 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-655mn"] Nov 22 05:45:51 crc kubenswrapper[4818]: I1122 05:45:51.236211 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:51 crc kubenswrapper[4818]: I1122 05:45:51.236483 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:45:51 crc kubenswrapper[4818]: I1122 05:45:51.264842 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:45:51 crc kubenswrapper[4818]: I1122 05:45:51.264915 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:45:51 crc kubenswrapper[4818]: I1122 05:45:51.320127 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a555d9aa-2ac5-4bf7-93de-63864d7fe9e0-catalog-content\") pod \"redhat-marketplace-655mn\" (UID: \"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0\") " pod="openshift-marketplace/redhat-marketplace-655mn" Nov 22 05:45:51 crc kubenswrapper[4818]: I1122 05:45:51.320222 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw9mc\" (UniqueName: \"kubernetes.io/projected/a555d9aa-2ac5-4bf7-93de-63864d7fe9e0-kube-api-access-bw9mc\") pod \"redhat-marketplace-655mn\" (UID: \"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0\") " pod="openshift-marketplace/redhat-marketplace-655mn" Nov 22 05:45:51 crc kubenswrapper[4818]: I1122 05:45:51.320625 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a555d9aa-2ac5-4bf7-93de-63864d7fe9e0-utilities\") pod \"redhat-marketplace-655mn\" (UID: \"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0\") " pod="openshift-marketplace/redhat-marketplace-655mn" Nov 22 05:45:51 crc kubenswrapper[4818]: I1122 05:45:51.422869 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw9mc\" (UniqueName: \"kubernetes.io/projected/a555d9aa-2ac5-4bf7-93de-63864d7fe9e0-kube-api-access-bw9mc\") pod \"redhat-marketplace-655mn\" (UID: \"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0\") " pod="openshift-marketplace/redhat-marketplace-655mn" Nov 22 05:45:51 crc kubenswrapper[4818]: I1122 05:45:51.423108 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a555d9aa-2ac5-4bf7-93de-63864d7fe9e0-utilities\") pod \"redhat-marketplace-655mn\" (UID: \"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0\") " pod="openshift-marketplace/redhat-marketplace-655mn" Nov 22 05:45:51 crc kubenswrapper[4818]: I1122 05:45:51.423304 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a555d9aa-2ac5-4bf7-93de-63864d7fe9e0-catalog-content\") pod \"redhat-marketplace-655mn\" (UID: \"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0\") " pod="openshift-marketplace/redhat-marketplace-655mn" Nov 22 05:45:51 crc kubenswrapper[4818]: I1122 05:45:51.426065 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a555d9aa-2ac5-4bf7-93de-63864d7fe9e0-utilities\") pod \"redhat-marketplace-655mn\" (UID: \"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0\") " pod="openshift-marketplace/redhat-marketplace-655mn" Nov 22 05:45:51 crc kubenswrapper[4818]: I1122 05:45:51.427517 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a555d9aa-2ac5-4bf7-93de-63864d7fe9e0-catalog-content\") pod \"redhat-marketplace-655mn\" (UID: \"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0\") " pod="openshift-marketplace/redhat-marketplace-655mn" Nov 22 05:45:51 crc kubenswrapper[4818]: I1122 05:45:51.472173 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw9mc\" (UniqueName: \"kubernetes.io/projected/a555d9aa-2ac5-4bf7-93de-63864d7fe9e0-kube-api-access-bw9mc\") pod \"redhat-marketplace-655mn\" (UID: \"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0\") " pod="openshift-marketplace/redhat-marketplace-655mn" Nov 22 05:45:51 crc kubenswrapper[4818]: I1122 05:45:51.495683 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-655mn" Nov 22 05:45:52 crc kubenswrapper[4818]: I1122 05:45:52.014359 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-655mn"] Nov 22 05:45:52 crc kubenswrapper[4818]: W1122 05:45:52.024227 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda555d9aa_2ac5_4bf7_93de_63864d7fe9e0.slice/crio-654ae8a08562c1a06f45c6e9716890e5de1a08a0c99b444f038d24f7af758074 WatchSource:0}: Error finding container 654ae8a08562c1a06f45c6e9716890e5de1a08a0c99b444f038d24f7af758074: Status 404 returned error can't find the container with id 654ae8a08562c1a06f45c6e9716890e5de1a08a0c99b444f038d24f7af758074 Nov 22 05:45:52 crc kubenswrapper[4818]: I1122 05:45:52.295727 4818 generic.go:334] "Generic (PLEG): container finished" podID="a555d9aa-2ac5-4bf7-93de-63864d7fe9e0" containerID="8d899477348741f8f9eff882fec4da6c783b25a8b661fafb6248060ba8758535" exitCode=0 Nov 22 05:45:52 crc kubenswrapper[4818]: I1122 05:45:52.310504 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-655mn" event={"ID":"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0","Type":"ContainerDied","Data":"8d899477348741f8f9eff882fec4da6c783b25a8b661fafb6248060ba8758535"} Nov 22 05:45:52 crc kubenswrapper[4818]: I1122 05:45:52.314712 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-655mn" event={"ID":"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0","Type":"ContainerStarted","Data":"654ae8a08562c1a06f45c6e9716890e5de1a08a0c99b444f038d24f7af758074"} Nov 22 05:45:52 crc kubenswrapper[4818]: E1122 05:45:52.339145 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3759d20_d679_42de_9a4e_f32305ec1ca0.slice/crio-aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3759d20_d679_42de_9a4e_f32305ec1ca0.slice/crio-conmon-aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514.scope\": RecentStats: unable to find data in memory cache]" Nov 22 05:45:53 crc kubenswrapper[4818]: I1122 05:45:53.313463 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-655mn" event={"ID":"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0","Type":"ContainerStarted","Data":"8548a993efaed33a954446f06681157fbb45ebd0d238e16eff9d94c9ef46a57b"} Nov 22 05:45:54 crc kubenswrapper[4818]: I1122 05:45:54.325713 4818 generic.go:334] "Generic (PLEG): container finished" podID="a555d9aa-2ac5-4bf7-93de-63864d7fe9e0" containerID="8548a993efaed33a954446f06681157fbb45ebd0d238e16eff9d94c9ef46a57b" exitCode=0 Nov 22 05:45:54 crc kubenswrapper[4818]: I1122 05:45:54.325766 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-655mn" event={"ID":"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0","Type":"ContainerDied","Data":"8548a993efaed33a954446f06681157fbb45ebd0d238e16eff9d94c9ef46a57b"} Nov 22 05:45:55 crc kubenswrapper[4818]: I1122 05:45:55.343402 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-655mn" event={"ID":"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0","Type":"ContainerStarted","Data":"7af190b3d41d6b54abd391ccf4e8910df92a5fb4f85b23f743084b162a64767f"} Nov 22 05:45:55 crc kubenswrapper[4818]: I1122 05:45:55.462493 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:55 crc kubenswrapper[4818]: I1122 05:45:55.462535 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:55 crc kubenswrapper[4818]: I1122 05:45:55.513716 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:55 crc kubenswrapper[4818]: I1122 05:45:55.540194 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-655mn" podStartSLOduration=2.117356043 podStartE2EDuration="4.54016988s" podCreationTimestamp="2025-11-22 05:45:51 +0000 UTC" firstStartedPulling="2025-11-22 05:45:52.297530669 +0000 UTC m=+3504.871947196" lastFinishedPulling="2025-11-22 05:45:54.720344496 +0000 UTC m=+3507.294761033" observedRunningTime="2025-11-22 05:45:55.370465496 +0000 UTC m=+3507.944882043" watchObservedRunningTime="2025-11-22 05:45:55.54016988 +0000 UTC m=+3508.114586417" Nov 22 05:45:55 crc kubenswrapper[4818]: I1122 05:45:55.547206 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:56 crc kubenswrapper[4818]: I1122 05:45:56.353453 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:56 crc kubenswrapper[4818]: I1122 05:45:56.354770 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:57 crc kubenswrapper[4818]: I1122 05:45:57.615989 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ksfkm" Nov 22 05:45:57 crc kubenswrapper[4818]: I1122 05:45:57.692701 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ksfkm" Nov 22 05:45:58 crc kubenswrapper[4818]: I1122 05:45:58.380763 4818 generic.go:334] "Generic (PLEG): container finished" podID="64f4e98d-2953-472c-852a-424b72c513a9" containerID="efd6a02edfb01d2064fc0e366c29accec0a100d9d2b56e5807f6c968832182b8" exitCode=0 Nov 22 05:45:58 crc kubenswrapper[4818]: I1122 05:45:58.381662 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-2xttz" event={"ID":"64f4e98d-2953-472c-852a-424b72c513a9","Type":"ContainerDied","Data":"efd6a02edfb01d2064fc0e366c29accec0a100d9d2b56e5807f6c968832182b8"} Nov 22 05:45:58 crc kubenswrapper[4818]: I1122 05:45:58.381967 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 05:45:58 crc kubenswrapper[4818]: I1122 05:45:58.381990 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 05:45:58 crc kubenswrapper[4818]: I1122 05:45:58.440744 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:58 crc kubenswrapper[4818]: I1122 05:45:58.530043 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ksfkm"] Nov 22 05:45:58 crc kubenswrapper[4818]: I1122 05:45:58.722319 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 22 05:45:59 crc kubenswrapper[4818]: I1122 05:45:59.393701 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ksfkm" podUID="e3759d20-d679-42de-9a4e-f32305ec1ca0" containerName="registry-server" containerID="cri-o://e499a2f0e3524f430fa8bb42729769b665dc5cb1795195e03f7620c112479094" gracePeriod=2 Nov 22 05:45:59 crc kubenswrapper[4818]: I1122 05:45:59.910700 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-2xttz" Nov 22 05:45:59 crc kubenswrapper[4818]: I1122 05:45:59.935984 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ksfkm" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.091735 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3759d20-d679-42de-9a4e-f32305ec1ca0-utilities\") pod \"e3759d20-d679-42de-9a4e-f32305ec1ca0\" (UID: \"e3759d20-d679-42de-9a4e-f32305ec1ca0\") " Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.091801 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64f4e98d-2953-472c-852a-424b72c513a9-config-data\") pod \"64f4e98d-2953-472c-852a-424b72c513a9\" (UID: \"64f4e98d-2953-472c-852a-424b72c513a9\") " Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.091885 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9xxb\" (UniqueName: \"kubernetes.io/projected/e3759d20-d679-42de-9a4e-f32305ec1ca0-kube-api-access-q9xxb\") pod \"e3759d20-d679-42de-9a4e-f32305ec1ca0\" (UID: \"e3759d20-d679-42de-9a4e-f32305ec1ca0\") " Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.091914 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3759d20-d679-42de-9a4e-f32305ec1ca0-catalog-content\") pod \"e3759d20-d679-42de-9a4e-f32305ec1ca0\" (UID: \"e3759d20-d679-42de-9a4e-f32305ec1ca0\") " Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.091988 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d58gq\" (UniqueName: \"kubernetes.io/projected/64f4e98d-2953-472c-852a-424b72c513a9-kube-api-access-d58gq\") pod \"64f4e98d-2953-472c-852a-424b72c513a9\" (UID: \"64f4e98d-2953-472c-852a-424b72c513a9\") " Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.092032 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/64f4e98d-2953-472c-852a-424b72c513a9-job-config-data\") pod \"64f4e98d-2953-472c-852a-424b72c513a9\" (UID: \"64f4e98d-2953-472c-852a-424b72c513a9\") " Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.092060 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64f4e98d-2953-472c-852a-424b72c513a9-combined-ca-bundle\") pod \"64f4e98d-2953-472c-852a-424b72c513a9\" (UID: \"64f4e98d-2953-472c-852a-424b72c513a9\") " Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.093839 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3759d20-d679-42de-9a4e-f32305ec1ca0-utilities" (OuterVolumeSpecName: "utilities") pod "e3759d20-d679-42de-9a4e-f32305ec1ca0" (UID: "e3759d20-d679-42de-9a4e-f32305ec1ca0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.100748 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3759d20-d679-42de-9a4e-f32305ec1ca0-kube-api-access-q9xxb" (OuterVolumeSpecName: "kube-api-access-q9xxb") pod "e3759d20-d679-42de-9a4e-f32305ec1ca0" (UID: "e3759d20-d679-42de-9a4e-f32305ec1ca0"). InnerVolumeSpecName "kube-api-access-q9xxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.100822 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64f4e98d-2953-472c-852a-424b72c513a9-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "64f4e98d-2953-472c-852a-424b72c513a9" (UID: "64f4e98d-2953-472c-852a-424b72c513a9"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.104483 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64f4e98d-2953-472c-852a-424b72c513a9-kube-api-access-d58gq" (OuterVolumeSpecName: "kube-api-access-d58gq") pod "64f4e98d-2953-472c-852a-424b72c513a9" (UID: "64f4e98d-2953-472c-852a-424b72c513a9"). InnerVolumeSpecName "kube-api-access-d58gq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.119422 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64f4e98d-2953-472c-852a-424b72c513a9-config-data" (OuterVolumeSpecName: "config-data") pod "64f4e98d-2953-472c-852a-424b72c513a9" (UID: "64f4e98d-2953-472c-852a-424b72c513a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.133374 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64f4e98d-2953-472c-852a-424b72c513a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "64f4e98d-2953-472c-852a-424b72c513a9" (UID: "64f4e98d-2953-472c-852a-424b72c513a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.155772 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3759d20-d679-42de-9a4e-f32305ec1ca0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e3759d20-d679-42de-9a4e-f32305ec1ca0" (UID: "e3759d20-d679-42de-9a4e-f32305ec1ca0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.194637 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9xxb\" (UniqueName: \"kubernetes.io/projected/e3759d20-d679-42de-9a4e-f32305ec1ca0-kube-api-access-q9xxb\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.194993 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3759d20-d679-42de-9a4e-f32305ec1ca0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.195076 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d58gq\" (UniqueName: \"kubernetes.io/projected/64f4e98d-2953-472c-852a-424b72c513a9-kube-api-access-d58gq\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.195138 4818 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/64f4e98d-2953-472c-852a-424b72c513a9-job-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.195391 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64f4e98d-2953-472c-852a-424b72c513a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.195464 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3759d20-d679-42de-9a4e-f32305ec1ca0-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.195518 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64f4e98d-2953-472c-852a-424b72c513a9-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.410390 4818 generic.go:334] "Generic (PLEG): container finished" podID="e3759d20-d679-42de-9a4e-f32305ec1ca0" containerID="e499a2f0e3524f430fa8bb42729769b665dc5cb1795195e03f7620c112479094" exitCode=0 Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.410485 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ksfkm" event={"ID":"e3759d20-d679-42de-9a4e-f32305ec1ca0","Type":"ContainerDied","Data":"e499a2f0e3524f430fa8bb42729769b665dc5cb1795195e03f7620c112479094"} Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.410540 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ksfkm" event={"ID":"e3759d20-d679-42de-9a4e-f32305ec1ca0","Type":"ContainerDied","Data":"111f708a9cacde7583bf2e26ada8a9d1460f6d13ff5ffd9b987c416bf415431a"} Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.410562 4818 scope.go:117] "RemoveContainer" containerID="e499a2f0e3524f430fa8bb42729769b665dc5cb1795195e03f7620c112479094" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.410772 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ksfkm" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.432050 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-2xttz" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.432164 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-2xttz" event={"ID":"64f4e98d-2953-472c-852a-424b72c513a9","Type":"ContainerDied","Data":"13bb3ccd64b4270bb8706864751f2ab55a499d5559887785e193a1dc1f688b57"} Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.432230 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13bb3ccd64b4270bb8706864751f2ab55a499d5559887785e193a1dc1f688b57" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.462566 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ksfkm"] Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.463705 4818 scope.go:117] "RemoveContainer" containerID="f41b189f96fd2137bd4914485c32775c0dfb95bb75ea890ad134e6edfdc260b4" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.471501 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ksfkm"] Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.484949 4818 scope.go:117] "RemoveContainer" containerID="aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.584135 4818 scope.go:117] "RemoveContainer" containerID="e499a2f0e3524f430fa8bb42729769b665dc5cb1795195e03f7620c112479094" Nov 22 05:46:00 crc kubenswrapper[4818]: E1122 05:46:00.585715 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e499a2f0e3524f430fa8bb42729769b665dc5cb1795195e03f7620c112479094\": container with ID starting with e499a2f0e3524f430fa8bb42729769b665dc5cb1795195e03f7620c112479094 not found: ID does not exist" containerID="e499a2f0e3524f430fa8bb42729769b665dc5cb1795195e03f7620c112479094" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.585747 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e499a2f0e3524f430fa8bb42729769b665dc5cb1795195e03f7620c112479094"} err="failed to get container status \"e499a2f0e3524f430fa8bb42729769b665dc5cb1795195e03f7620c112479094\": rpc error: code = NotFound desc = could not find container \"e499a2f0e3524f430fa8bb42729769b665dc5cb1795195e03f7620c112479094\": container with ID starting with e499a2f0e3524f430fa8bb42729769b665dc5cb1795195e03f7620c112479094 not found: ID does not exist" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.585767 4818 scope.go:117] "RemoveContainer" containerID="f41b189f96fd2137bd4914485c32775c0dfb95bb75ea890ad134e6edfdc260b4" Nov 22 05:46:00 crc kubenswrapper[4818]: E1122 05:46:00.589601 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f41b189f96fd2137bd4914485c32775c0dfb95bb75ea890ad134e6edfdc260b4\": container with ID starting with f41b189f96fd2137bd4914485c32775c0dfb95bb75ea890ad134e6edfdc260b4 not found: ID does not exist" containerID="f41b189f96fd2137bd4914485c32775c0dfb95bb75ea890ad134e6edfdc260b4" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.589637 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f41b189f96fd2137bd4914485c32775c0dfb95bb75ea890ad134e6edfdc260b4"} err="failed to get container status \"f41b189f96fd2137bd4914485c32775c0dfb95bb75ea890ad134e6edfdc260b4\": rpc error: code = NotFound desc = could not find container \"f41b189f96fd2137bd4914485c32775c0dfb95bb75ea890ad134e6edfdc260b4\": container with ID starting with f41b189f96fd2137bd4914485c32775c0dfb95bb75ea890ad134e6edfdc260b4 not found: ID does not exist" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.589660 4818 scope.go:117] "RemoveContainer" containerID="aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514" Nov 22 05:46:00 crc kubenswrapper[4818]: E1122 05:46:00.590157 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514\": container with ID starting with aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514 not found: ID does not exist" containerID="aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.590395 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514"} err="failed to get container status \"aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514\": rpc error: code = NotFound desc = could not find container \"aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514\": container with ID starting with aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514 not found: ID does not exist" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.667693 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Nov 22 05:46:00 crc kubenswrapper[4818]: E1122 05:46:00.668290 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3759d20-d679-42de-9a4e-f32305ec1ca0" containerName="registry-server" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.678611 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3759d20-d679-42de-9a4e-f32305ec1ca0" containerName="registry-server" Nov 22 05:46:00 crc kubenswrapper[4818]: E1122 05:46:00.678746 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3759d20-d679-42de-9a4e-f32305ec1ca0" containerName="extract-content" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.678799 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3759d20-d679-42de-9a4e-f32305ec1ca0" containerName="extract-content" Nov 22 05:46:00 crc kubenswrapper[4818]: E1122 05:46:00.678875 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64f4e98d-2953-472c-852a-424b72c513a9" containerName="manila-db-sync" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.678960 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="64f4e98d-2953-472c-852a-424b72c513a9" containerName="manila-db-sync" Nov 22 05:46:00 crc kubenswrapper[4818]: E1122 05:46:00.679047 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3759d20-d679-42de-9a4e-f32305ec1ca0" containerName="extract-utilities" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.679101 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3759d20-d679-42de-9a4e-f32305ec1ca0" containerName="extract-utilities" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.679504 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="64f4e98d-2953-472c-852a-424b72c513a9" containerName="manila-db-sync" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.679583 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3759d20-d679-42de-9a4e-f32305ec1ca0" containerName="registry-server" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.680652 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.687933 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.688161 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-5jfbb" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.688299 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.688369 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.700934 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.757559 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d192104-a1b7-487b-bfda-72908d31075c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.757659 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.757805 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-config-data\") pod \"manila-scheduler-0\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.757873 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8cd9\" (UniqueName: \"kubernetes.io/projected/8d192104-a1b7-487b-bfda-72908d31075c-kube-api-access-j8cd9\") pod \"manila-scheduler-0\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.757903 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-scripts\") pod \"manila-scheduler-0\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.757936 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.780442 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.798911 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.801972 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.822760 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.859334 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8cd9\" (UniqueName: \"kubernetes.io/projected/8d192104-a1b7-487b-bfda-72908d31075c-kube-api-access-j8cd9\") pod \"manila-scheduler-0\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.859595 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-scripts\") pod \"manila-scheduler-0\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.859676 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.859802 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d192104-a1b7-487b-bfda-72908d31075c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.859897 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.860098 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-config-data\") pod \"manila-scheduler-0\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.861353 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76b5fdb995-zl4gj"] Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.864734 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d192104-a1b7-487b-bfda-72908d31075c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.867983 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.869842 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.870027 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-scripts\") pod \"manila-scheduler-0\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.870560 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-config-data\") pod \"manila-scheduler-0\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.876027 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.900602 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8cd9\" (UniqueName: \"kubernetes.io/projected/8d192104-a1b7-487b-bfda-72908d31075c-kube-api-access-j8cd9\") pod \"manila-scheduler-0\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.903843 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76b5fdb995-zl4gj"] Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.968527 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.968959 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/46554362-2c6d-45b0-b725-e32873db8d22-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.969016 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-scripts\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.969171 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46554362-2c6d-45b0-b725-e32873db8d22-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.969195 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.969227 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gbdz\" (UniqueName: \"kubernetes.io/projected/46554362-2c6d-45b0-b725-e32873db8d22-kube-api-access-9gbdz\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.970195 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.971145 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/46554362-2c6d-45b0-b725-e32873db8d22-ceph\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.971192 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.971227 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-config-data\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:00 crc kubenswrapper[4818]: I1122 05:46:00.973231 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.020720 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.033069 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.072648 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b-ovsdbserver-sb\") pod \"dnsmasq-dns-76b5fdb995-zl4gj\" (UID: \"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b\") " pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.072688 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/46554362-2c6d-45b0-b725-e32873db8d22-ceph\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.072735 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.072756 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-config-data-custom\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.072775 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-config-data\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.072821 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.072847 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs8rz\" (UniqueName: \"kubernetes.io/projected/cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b-kube-api-access-bs8rz\") pod \"dnsmasq-dns-76b5fdb995-zl4gj\" (UID: \"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b\") " pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.072870 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-config-data\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.072893 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b-openstack-edpm-ipam\") pod \"dnsmasq-dns-76b5fdb995-zl4gj\" (UID: \"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b\") " pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.072949 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/46554362-2c6d-45b0-b725-e32873db8d22-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.072974 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-scripts\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.072993 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/860fd889-3c2c-4202-b383-5491ca2d4e9d-etc-machine-id\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.073010 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/860fd889-3c2c-4202-b383-5491ca2d4e9d-logs\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.073032 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-scripts\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.073061 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b-ovsdbserver-nb\") pod \"dnsmasq-dns-76b5fdb995-zl4gj\" (UID: \"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b\") " pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.073100 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46554362-2c6d-45b0-b725-e32873db8d22-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.073117 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.073134 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phbvm\" (UniqueName: \"kubernetes.io/projected/860fd889-3c2c-4202-b383-5491ca2d4e9d-kube-api-access-phbvm\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.073151 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b-config\") pod \"dnsmasq-dns-76b5fdb995-zl4gj\" (UID: \"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b\") " pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.073167 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gbdz\" (UniqueName: \"kubernetes.io/projected/46554362-2c6d-45b0-b725-e32873db8d22-kube-api-access-9gbdz\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.073187 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b-dns-svc\") pod \"dnsmasq-dns-76b5fdb995-zl4gj\" (UID: \"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b\") " pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.074372 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46554362-2c6d-45b0-b725-e32873db8d22-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.074546 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/46554362-2c6d-45b0-b725-e32873db8d22-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.078473 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-config-data\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.078533 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/46554362-2c6d-45b0-b725-e32873db8d22-ceph\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.088917 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.090692 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.093739 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gbdz\" (UniqueName: \"kubernetes.io/projected/46554362-2c6d-45b0-b725-e32873db8d22-kube-api-access-9gbdz\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.097611 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-scripts\") pod \"manila-share-share1-0\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.103399 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7764859bb4-tnzkk" podUID="60b9da36-1281-4cc5-a9ac-a2df2e4d9638" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.242:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.242:8443: connect: connection refused" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.117957 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.175249 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.175563 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs8rz\" (UniqueName: \"kubernetes.io/projected/cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b-kube-api-access-bs8rz\") pod \"dnsmasq-dns-76b5fdb995-zl4gj\" (UID: \"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b\") " pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.175589 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-config-data\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.175616 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b-openstack-edpm-ipam\") pod \"dnsmasq-dns-76b5fdb995-zl4gj\" (UID: \"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b\") " pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.175671 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/860fd889-3c2c-4202-b383-5491ca2d4e9d-etc-machine-id\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.175687 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/860fd889-3c2c-4202-b383-5491ca2d4e9d-logs\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.175709 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-scripts\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.175734 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b-ovsdbserver-nb\") pod \"dnsmasq-dns-76b5fdb995-zl4gj\" (UID: \"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b\") " pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.175773 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phbvm\" (UniqueName: \"kubernetes.io/projected/860fd889-3c2c-4202-b383-5491ca2d4e9d-kube-api-access-phbvm\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.175789 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b-config\") pod \"dnsmasq-dns-76b5fdb995-zl4gj\" (UID: \"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b\") " pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.175807 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b-dns-svc\") pod \"dnsmasq-dns-76b5fdb995-zl4gj\" (UID: \"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b\") " pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.175831 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b-ovsdbserver-sb\") pod \"dnsmasq-dns-76b5fdb995-zl4gj\" (UID: \"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b\") " pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.175856 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-config-data-custom\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.178183 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b-openstack-edpm-ipam\") pod \"dnsmasq-dns-76b5fdb995-zl4gj\" (UID: \"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b\") " pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.178732 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b-dns-svc\") pod \"dnsmasq-dns-76b5fdb995-zl4gj\" (UID: \"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b\") " pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.178802 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b-config\") pod \"dnsmasq-dns-76b5fdb995-zl4gj\" (UID: \"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b\") " pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.179337 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b-ovsdbserver-sb\") pod \"dnsmasq-dns-76b5fdb995-zl4gj\" (UID: \"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b\") " pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.179421 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b-ovsdbserver-nb\") pod \"dnsmasq-dns-76b5fdb995-zl4gj\" (UID: \"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b\") " pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.179463 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/860fd889-3c2c-4202-b383-5491ca2d4e9d-etc-machine-id\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.179507 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.179668 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/860fd889-3c2c-4202-b383-5491ca2d4e9d-logs\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.179981 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-scripts\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.180786 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-config-data-custom\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.182246 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-config-data\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.197905 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phbvm\" (UniqueName: \"kubernetes.io/projected/860fd889-3c2c-4202-b383-5491ca2d4e9d-kube-api-access-phbvm\") pod \"manila-api-0\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.204055 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs8rz\" (UniqueName: \"kubernetes.io/projected/cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b-kube-api-access-bs8rz\") pod \"dnsmasq-dns-76b5fdb995-zl4gj\" (UID: \"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b\") " pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.242404 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-86757bfcb-s9bbk" podUID="34f2acb8-906b-4520-86c2-bb603a964489" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.243:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.243:8443: connect: connection refused" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.317765 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.328124 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.497640 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-655mn" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.497888 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-655mn" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.666662 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-655mn" Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.836014 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 22 05:46:01 crc kubenswrapper[4818]: W1122 05:46:01.857812 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d192104_a1b7_487b_bfda_72908d31075c.slice/crio-623fe0080210daf12ec0904c4c93aec009ab8f41de24e7b5c56bab49321c0ed3 WatchSource:0}: Error finding container 623fe0080210daf12ec0904c4c93aec009ab8f41de24e7b5c56bab49321c0ed3: Status 404 returned error can't find the container with id 623fe0080210daf12ec0904c4c93aec009ab8f41de24e7b5c56bab49321c0ed3 Nov 22 05:46:01 crc kubenswrapper[4818]: I1122 05:46:01.988503 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 22 05:46:01 crc kubenswrapper[4818]: W1122 05:46:01.996583 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46554362_2c6d_45b0_b725_e32873db8d22.slice/crio-aa5028bef55a81fc8b30918c1cf15a51fb39367c70864f9390e7c56aaf26b8a1 WatchSource:0}: Error finding container aa5028bef55a81fc8b30918c1cf15a51fb39367c70864f9390e7c56aaf26b8a1: Status 404 returned error can't find the container with id aa5028bef55a81fc8b30918c1cf15a51fb39367c70864f9390e7c56aaf26b8a1 Nov 22 05:46:02 crc kubenswrapper[4818]: I1122 05:46:02.227243 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76b5fdb995-zl4gj"] Nov 22 05:46:02 crc kubenswrapper[4818]: I1122 05:46:02.300208 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3759d20-d679-42de-9a4e-f32305ec1ca0" path="/var/lib/kubelet/pods/e3759d20-d679-42de-9a4e-f32305ec1ca0/volumes" Nov 22 05:46:02 crc kubenswrapper[4818]: I1122 05:46:02.321768 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 22 05:46:02 crc kubenswrapper[4818]: I1122 05:46:02.466988 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"46554362-2c6d-45b0-b725-e32873db8d22","Type":"ContainerStarted","Data":"aa5028bef55a81fc8b30918c1cf15a51fb39367c70864f9390e7c56aaf26b8a1"} Nov 22 05:46:02 crc kubenswrapper[4818]: I1122 05:46:02.469363 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" event={"ID":"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b","Type":"ContainerStarted","Data":"a8b6a181c2d7757a45ce0bf239ec364920b17086eb88ab44c1988be622b45fe0"} Nov 22 05:46:02 crc kubenswrapper[4818]: I1122 05:46:02.476428 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8d192104-a1b7-487b-bfda-72908d31075c","Type":"ContainerStarted","Data":"623fe0080210daf12ec0904c4c93aec009ab8f41de24e7b5c56bab49321c0ed3"} Nov 22 05:46:02 crc kubenswrapper[4818]: I1122 05:46:02.481150 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"860fd889-3c2c-4202-b383-5491ca2d4e9d","Type":"ContainerStarted","Data":"fd4f2017ec311d12121f7e0877432ab867217da8ee4c6911a019c97d3e961135"} Nov 22 05:46:02 crc kubenswrapper[4818]: I1122 05:46:02.559835 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-655mn" Nov 22 05:46:02 crc kubenswrapper[4818]: E1122 05:46:02.650214 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3759d20_d679_42de_9a4e_f32305ec1ca0.slice/crio-conmon-aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3759d20_d679_42de_9a4e_f32305ec1ca0.slice/crio-aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514.scope\": RecentStats: unable to find data in memory cache]" Nov 22 05:46:02 crc kubenswrapper[4818]: I1122 05:46:02.934707 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-655mn"] Nov 22 05:46:03 crc kubenswrapper[4818]: I1122 05:46:03.503490 4818 generic.go:334] "Generic (PLEG): container finished" podID="cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b" containerID="98afc76b76d90f81a98a3be43945bb58d9d279361dfde4c6a4197199b2243947" exitCode=0 Nov 22 05:46:03 crc kubenswrapper[4818]: I1122 05:46:03.504244 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" event={"ID":"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b","Type":"ContainerDied","Data":"98afc76b76d90f81a98a3be43945bb58d9d279361dfde4c6a4197199b2243947"} Nov 22 05:46:03 crc kubenswrapper[4818]: I1122 05:46:03.547598 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8d192104-a1b7-487b-bfda-72908d31075c","Type":"ContainerStarted","Data":"07b9628488431caba14b07b4eeb098576d3f811d9caf39f4ffdfcd3351ac3001"} Nov 22 05:46:03 crc kubenswrapper[4818]: I1122 05:46:03.554197 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"860fd889-3c2c-4202-b383-5491ca2d4e9d","Type":"ContainerStarted","Data":"9051d50c2a6c41c87399558b71bdf13cf73621ff40b92ab9e2ec55df5591e3a0"} Nov 22 05:46:03 crc kubenswrapper[4818]: I1122 05:46:03.883587 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Nov 22 05:46:04 crc kubenswrapper[4818]: I1122 05:46:04.565546 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"860fd889-3c2c-4202-b383-5491ca2d4e9d","Type":"ContainerStarted","Data":"fc1e3435e58e1270e767da8cf63b8cbdd4e1521ade2f9f25569aeffab472b53f"} Nov 22 05:46:04 crc kubenswrapper[4818]: I1122 05:46:04.566214 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Nov 22 05:46:04 crc kubenswrapper[4818]: I1122 05:46:04.569329 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" event={"ID":"cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b","Type":"ContainerStarted","Data":"2ace2e0ee96c3f433b46130db25240db3a233aa7806556a46fa3174291788c40"} Nov 22 05:46:04 crc kubenswrapper[4818]: I1122 05:46:04.569508 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:04 crc kubenswrapper[4818]: I1122 05:46:04.575468 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-655mn" podUID="a555d9aa-2ac5-4bf7-93de-63864d7fe9e0" containerName="registry-server" containerID="cri-o://7af190b3d41d6b54abd391ccf4e8910df92a5fb4f85b23f743084b162a64767f" gracePeriod=2 Nov 22 05:46:04 crc kubenswrapper[4818]: I1122 05:46:04.576203 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8d192104-a1b7-487b-bfda-72908d31075c","Type":"ContainerStarted","Data":"81fe080118a54161335c509815cd015f364ff2ab8dc85613816907f3047783f2"} Nov 22 05:46:04 crc kubenswrapper[4818]: I1122 05:46:04.587778 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=4.587761925 podStartE2EDuration="4.587761925s" podCreationTimestamp="2025-11-22 05:46:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:46:04.582957958 +0000 UTC m=+3517.157374505" watchObservedRunningTime="2025-11-22 05:46:04.587761925 +0000 UTC m=+3517.162178442" Nov 22 05:46:04 crc kubenswrapper[4818]: I1122 05:46:04.604520 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.875511518 podStartE2EDuration="4.604505892s" podCreationTimestamp="2025-11-22 05:46:00 +0000 UTC" firstStartedPulling="2025-11-22 05:46:01.868768324 +0000 UTC m=+3514.443184851" lastFinishedPulling="2025-11-22 05:46:02.597762708 +0000 UTC m=+3515.172179225" observedRunningTime="2025-11-22 05:46:04.602413667 +0000 UTC m=+3517.176830194" watchObservedRunningTime="2025-11-22 05:46:04.604505892 +0000 UTC m=+3517.178922419" Nov 22 05:46:04 crc kubenswrapper[4818]: I1122 05:46:04.624272 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" podStartSLOduration=4.624242218 podStartE2EDuration="4.624242218s" podCreationTimestamp="2025-11-22 05:46:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:46:04.617750045 +0000 UTC m=+3517.192166572" watchObservedRunningTime="2025-11-22 05:46:04.624242218 +0000 UTC m=+3517.198658735" Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.162334 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-655mn" Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.174986 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a555d9aa-2ac5-4bf7-93de-63864d7fe9e0-catalog-content\") pod \"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0\" (UID: \"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0\") " Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.175050 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a555d9aa-2ac5-4bf7-93de-63864d7fe9e0-utilities\") pod \"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0\" (UID: \"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0\") " Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.175332 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw9mc\" (UniqueName: \"kubernetes.io/projected/a555d9aa-2ac5-4bf7-93de-63864d7fe9e0-kube-api-access-bw9mc\") pod \"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0\" (UID: \"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0\") " Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.184787 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a555d9aa-2ac5-4bf7-93de-63864d7fe9e0-utilities" (OuterVolumeSpecName: "utilities") pod "a555d9aa-2ac5-4bf7-93de-63864d7fe9e0" (UID: "a555d9aa-2ac5-4bf7-93de-63864d7fe9e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.200912 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a555d9aa-2ac5-4bf7-93de-63864d7fe9e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a555d9aa-2ac5-4bf7-93de-63864d7fe9e0" (UID: "a555d9aa-2ac5-4bf7-93de-63864d7fe9e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.218768 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a555d9aa-2ac5-4bf7-93de-63864d7fe9e0-kube-api-access-bw9mc" (OuterVolumeSpecName: "kube-api-access-bw9mc") pod "a555d9aa-2ac5-4bf7-93de-63864d7fe9e0" (UID: "a555d9aa-2ac5-4bf7-93de-63864d7fe9e0"). InnerVolumeSpecName "kube-api-access-bw9mc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.278007 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw9mc\" (UniqueName: \"kubernetes.io/projected/a555d9aa-2ac5-4bf7-93de-63864d7fe9e0-kube-api-access-bw9mc\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.278041 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a555d9aa-2ac5-4bf7-93de-63864d7fe9e0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.278050 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a555d9aa-2ac5-4bf7-93de-63864d7fe9e0-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.623523 4818 generic.go:334] "Generic (PLEG): container finished" podID="a555d9aa-2ac5-4bf7-93de-63864d7fe9e0" containerID="7af190b3d41d6b54abd391ccf4e8910df92a5fb4f85b23f743084b162a64767f" exitCode=0 Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.623587 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-655mn" event={"ID":"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0","Type":"ContainerDied","Data":"7af190b3d41d6b54abd391ccf4e8910df92a5fb4f85b23f743084b162a64767f"} Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.623657 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-655mn" event={"ID":"a555d9aa-2ac5-4bf7-93de-63864d7fe9e0","Type":"ContainerDied","Data":"654ae8a08562c1a06f45c6e9716890e5de1a08a0c99b444f038d24f7af758074"} Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.623600 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-655mn" Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.623676 4818 scope.go:117] "RemoveContainer" containerID="7af190b3d41d6b54abd391ccf4e8910df92a5fb4f85b23f743084b162a64767f" Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.624033 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="860fd889-3c2c-4202-b383-5491ca2d4e9d" containerName="manila-api-log" containerID="cri-o://9051d50c2a6c41c87399558b71bdf13cf73621ff40b92ab9e2ec55df5591e3a0" gracePeriod=30 Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.624141 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="860fd889-3c2c-4202-b383-5491ca2d4e9d" containerName="manila-api" containerID="cri-o://fc1e3435e58e1270e767da8cf63b8cbdd4e1521ade2f9f25569aeffab472b53f" gracePeriod=30 Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.663856 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-655mn"] Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.672697 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-655mn"] Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.695271 4818 scope.go:117] "RemoveContainer" containerID="8548a993efaed33a954446f06681157fbb45ebd0d238e16eff9d94c9ef46a57b" Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.715460 4818 scope.go:117] "RemoveContainer" containerID="8d899477348741f8f9eff882fec4da6c783b25a8b661fafb6248060ba8758535" Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.925316 4818 scope.go:117] "RemoveContainer" containerID="7af190b3d41d6b54abd391ccf4e8910df92a5fb4f85b23f743084b162a64767f" Nov 22 05:46:05 crc kubenswrapper[4818]: E1122 05:46:05.926874 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7af190b3d41d6b54abd391ccf4e8910df92a5fb4f85b23f743084b162a64767f\": container with ID starting with 7af190b3d41d6b54abd391ccf4e8910df92a5fb4f85b23f743084b162a64767f not found: ID does not exist" containerID="7af190b3d41d6b54abd391ccf4e8910df92a5fb4f85b23f743084b162a64767f" Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.926907 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7af190b3d41d6b54abd391ccf4e8910df92a5fb4f85b23f743084b162a64767f"} err="failed to get container status \"7af190b3d41d6b54abd391ccf4e8910df92a5fb4f85b23f743084b162a64767f\": rpc error: code = NotFound desc = could not find container \"7af190b3d41d6b54abd391ccf4e8910df92a5fb4f85b23f743084b162a64767f\": container with ID starting with 7af190b3d41d6b54abd391ccf4e8910df92a5fb4f85b23f743084b162a64767f not found: ID does not exist" Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.926928 4818 scope.go:117] "RemoveContainer" containerID="8548a993efaed33a954446f06681157fbb45ebd0d238e16eff9d94c9ef46a57b" Nov 22 05:46:05 crc kubenswrapper[4818]: E1122 05:46:05.927689 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8548a993efaed33a954446f06681157fbb45ebd0d238e16eff9d94c9ef46a57b\": container with ID starting with 8548a993efaed33a954446f06681157fbb45ebd0d238e16eff9d94c9ef46a57b not found: ID does not exist" containerID="8548a993efaed33a954446f06681157fbb45ebd0d238e16eff9d94c9ef46a57b" Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.927717 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8548a993efaed33a954446f06681157fbb45ebd0d238e16eff9d94c9ef46a57b"} err="failed to get container status \"8548a993efaed33a954446f06681157fbb45ebd0d238e16eff9d94c9ef46a57b\": rpc error: code = NotFound desc = could not find container \"8548a993efaed33a954446f06681157fbb45ebd0d238e16eff9d94c9ef46a57b\": container with ID starting with 8548a993efaed33a954446f06681157fbb45ebd0d238e16eff9d94c9ef46a57b not found: ID does not exist" Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.927735 4818 scope.go:117] "RemoveContainer" containerID="8d899477348741f8f9eff882fec4da6c783b25a8b661fafb6248060ba8758535" Nov 22 05:46:05 crc kubenswrapper[4818]: E1122 05:46:05.927955 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d899477348741f8f9eff882fec4da6c783b25a8b661fafb6248060ba8758535\": container with ID starting with 8d899477348741f8f9eff882fec4da6c783b25a8b661fafb6248060ba8758535 not found: ID does not exist" containerID="8d899477348741f8f9eff882fec4da6c783b25a8b661fafb6248060ba8758535" Nov 22 05:46:05 crc kubenswrapper[4818]: I1122 05:46:05.927976 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d899477348741f8f9eff882fec4da6c783b25a8b661fafb6248060ba8758535"} err="failed to get container status \"8d899477348741f8f9eff882fec4da6c783b25a8b661fafb6248060ba8758535\": rpc error: code = NotFound desc = could not find container \"8d899477348741f8f9eff882fec4da6c783b25a8b661fafb6248060ba8758535\": container with ID starting with 8d899477348741f8f9eff882fec4da6c783b25a8b661fafb6248060ba8758535 not found: ID does not exist" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.247807 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.309623 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a555d9aa-2ac5-4bf7-93de-63864d7fe9e0" path="/var/lib/kubelet/pods/a555d9aa-2ac5-4bf7-93de-63864d7fe9e0/volumes" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.408903 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phbvm\" (UniqueName: \"kubernetes.io/projected/860fd889-3c2c-4202-b383-5491ca2d4e9d-kube-api-access-phbvm\") pod \"860fd889-3c2c-4202-b383-5491ca2d4e9d\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.408961 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-combined-ca-bundle\") pod \"860fd889-3c2c-4202-b383-5491ca2d4e9d\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.409038 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/860fd889-3c2c-4202-b383-5491ca2d4e9d-logs\") pod \"860fd889-3c2c-4202-b383-5491ca2d4e9d\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.409083 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/860fd889-3c2c-4202-b383-5491ca2d4e9d-etc-machine-id\") pod \"860fd889-3c2c-4202-b383-5491ca2d4e9d\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.409102 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-config-data\") pod \"860fd889-3c2c-4202-b383-5491ca2d4e9d\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.409119 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-scripts\") pod \"860fd889-3c2c-4202-b383-5491ca2d4e9d\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.409191 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-config-data-custom\") pod \"860fd889-3c2c-4202-b383-5491ca2d4e9d\" (UID: \"860fd889-3c2c-4202-b383-5491ca2d4e9d\") " Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.409648 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/860fd889-3c2c-4202-b383-5491ca2d4e9d-logs" (OuterVolumeSpecName: "logs") pod "860fd889-3c2c-4202-b383-5491ca2d4e9d" (UID: "860fd889-3c2c-4202-b383-5491ca2d4e9d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.412435 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/860fd889-3c2c-4202-b383-5491ca2d4e9d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "860fd889-3c2c-4202-b383-5491ca2d4e9d" (UID: "860fd889-3c2c-4202-b383-5491ca2d4e9d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.419466 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-scripts" (OuterVolumeSpecName: "scripts") pod "860fd889-3c2c-4202-b383-5491ca2d4e9d" (UID: "860fd889-3c2c-4202-b383-5491ca2d4e9d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.419487 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "860fd889-3c2c-4202-b383-5491ca2d4e9d" (UID: "860fd889-3c2c-4202-b383-5491ca2d4e9d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.419469 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/860fd889-3c2c-4202-b383-5491ca2d4e9d-kube-api-access-phbvm" (OuterVolumeSpecName: "kube-api-access-phbvm") pod "860fd889-3c2c-4202-b383-5491ca2d4e9d" (UID: "860fd889-3c2c-4202-b383-5491ca2d4e9d"). InnerVolumeSpecName "kube-api-access-phbvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.470153 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "860fd889-3c2c-4202-b383-5491ca2d4e9d" (UID: "860fd889-3c2c-4202-b383-5491ca2d4e9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.480148 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-config-data" (OuterVolumeSpecName: "config-data") pod "860fd889-3c2c-4202-b383-5491ca2d4e9d" (UID: "860fd889-3c2c-4202-b383-5491ca2d4e9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.512560 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.512585 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/860fd889-3c2c-4202-b383-5491ca2d4e9d-logs\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.512596 4818 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/860fd889-3c2c-4202-b383-5491ca2d4e9d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.512606 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.512614 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.512623 4818 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/860fd889-3c2c-4202-b383-5491ca2d4e9d-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.512632 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phbvm\" (UniqueName: \"kubernetes.io/projected/860fd889-3c2c-4202-b383-5491ca2d4e9d-kube-api-access-phbvm\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.640704 4818 generic.go:334] "Generic (PLEG): container finished" podID="860fd889-3c2c-4202-b383-5491ca2d4e9d" containerID="fc1e3435e58e1270e767da8cf63b8cbdd4e1521ade2f9f25569aeffab472b53f" exitCode=0 Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.640741 4818 generic.go:334] "Generic (PLEG): container finished" podID="860fd889-3c2c-4202-b383-5491ca2d4e9d" containerID="9051d50c2a6c41c87399558b71bdf13cf73621ff40b92ab9e2ec55df5591e3a0" exitCode=143 Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.640844 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.640839 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"860fd889-3c2c-4202-b383-5491ca2d4e9d","Type":"ContainerDied","Data":"fc1e3435e58e1270e767da8cf63b8cbdd4e1521ade2f9f25569aeffab472b53f"} Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.640882 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"860fd889-3c2c-4202-b383-5491ca2d4e9d","Type":"ContainerDied","Data":"9051d50c2a6c41c87399558b71bdf13cf73621ff40b92ab9e2ec55df5591e3a0"} Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.640893 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"860fd889-3c2c-4202-b383-5491ca2d4e9d","Type":"ContainerDied","Data":"fd4f2017ec311d12121f7e0877432ab867217da8ee4c6911a019c97d3e961135"} Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.640909 4818 scope.go:117] "RemoveContainer" containerID="fc1e3435e58e1270e767da8cf63b8cbdd4e1521ade2f9f25569aeffab472b53f" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.673409 4818 scope.go:117] "RemoveContainer" containerID="9051d50c2a6c41c87399558b71bdf13cf73621ff40b92ab9e2ec55df5591e3a0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.683923 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.700789 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.701600 4818 scope.go:117] "RemoveContainer" containerID="fc1e3435e58e1270e767da8cf63b8cbdd4e1521ade2f9f25569aeffab472b53f" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.710164 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Nov 22 05:46:06 crc kubenswrapper[4818]: E1122 05:46:06.710537 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="860fd889-3c2c-4202-b383-5491ca2d4e9d" containerName="manila-api" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.710548 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="860fd889-3c2c-4202-b383-5491ca2d4e9d" containerName="manila-api" Nov 22 05:46:06 crc kubenswrapper[4818]: E1122 05:46:06.710572 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a555d9aa-2ac5-4bf7-93de-63864d7fe9e0" containerName="extract-content" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.710577 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a555d9aa-2ac5-4bf7-93de-63864d7fe9e0" containerName="extract-content" Nov 22 05:46:06 crc kubenswrapper[4818]: E1122 05:46:06.710596 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a555d9aa-2ac5-4bf7-93de-63864d7fe9e0" containerName="extract-utilities" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.710602 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a555d9aa-2ac5-4bf7-93de-63864d7fe9e0" containerName="extract-utilities" Nov 22 05:46:06 crc kubenswrapper[4818]: E1122 05:46:06.710611 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="860fd889-3c2c-4202-b383-5491ca2d4e9d" containerName="manila-api-log" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.710618 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="860fd889-3c2c-4202-b383-5491ca2d4e9d" containerName="manila-api-log" Nov 22 05:46:06 crc kubenswrapper[4818]: E1122 05:46:06.710629 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a555d9aa-2ac5-4bf7-93de-63864d7fe9e0" containerName="registry-server" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.710635 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a555d9aa-2ac5-4bf7-93de-63864d7fe9e0" containerName="registry-server" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.710791 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="860fd889-3c2c-4202-b383-5491ca2d4e9d" containerName="manila-api-log" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.710808 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="860fd889-3c2c-4202-b383-5491ca2d4e9d" containerName="manila-api" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.710843 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a555d9aa-2ac5-4bf7-93de-63864d7fe9e0" containerName="registry-server" Nov 22 05:46:06 crc kubenswrapper[4818]: E1122 05:46:06.711454 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc1e3435e58e1270e767da8cf63b8cbdd4e1521ade2f9f25569aeffab472b53f\": container with ID starting with fc1e3435e58e1270e767da8cf63b8cbdd4e1521ade2f9f25569aeffab472b53f not found: ID does not exist" containerID="fc1e3435e58e1270e767da8cf63b8cbdd4e1521ade2f9f25569aeffab472b53f" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.711495 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc1e3435e58e1270e767da8cf63b8cbdd4e1521ade2f9f25569aeffab472b53f"} err="failed to get container status \"fc1e3435e58e1270e767da8cf63b8cbdd4e1521ade2f9f25569aeffab472b53f\": rpc error: code = NotFound desc = could not find container \"fc1e3435e58e1270e767da8cf63b8cbdd4e1521ade2f9f25569aeffab472b53f\": container with ID starting with fc1e3435e58e1270e767da8cf63b8cbdd4e1521ade2f9f25569aeffab472b53f not found: ID does not exist" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.711520 4818 scope.go:117] "RemoveContainer" containerID="9051d50c2a6c41c87399558b71bdf13cf73621ff40b92ab9e2ec55df5591e3a0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.711773 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: E1122 05:46:06.723602 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9051d50c2a6c41c87399558b71bdf13cf73621ff40b92ab9e2ec55df5591e3a0\": container with ID starting with 9051d50c2a6c41c87399558b71bdf13cf73621ff40b92ab9e2ec55df5591e3a0 not found: ID does not exist" containerID="9051d50c2a6c41c87399558b71bdf13cf73621ff40b92ab9e2ec55df5591e3a0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.723655 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9051d50c2a6c41c87399558b71bdf13cf73621ff40b92ab9e2ec55df5591e3a0"} err="failed to get container status \"9051d50c2a6c41c87399558b71bdf13cf73621ff40b92ab9e2ec55df5591e3a0\": rpc error: code = NotFound desc = could not find container \"9051d50c2a6c41c87399558b71bdf13cf73621ff40b92ab9e2ec55df5591e3a0\": container with ID starting with 9051d50c2a6c41c87399558b71bdf13cf73621ff40b92ab9e2ec55df5591e3a0 not found: ID does not exist" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.723682 4818 scope.go:117] "RemoveContainer" containerID="fc1e3435e58e1270e767da8cf63b8cbdd4e1521ade2f9f25569aeffab472b53f" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.725175 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc1e3435e58e1270e767da8cf63b8cbdd4e1521ade2f9f25569aeffab472b53f"} err="failed to get container status \"fc1e3435e58e1270e767da8cf63b8cbdd4e1521ade2f9f25569aeffab472b53f\": rpc error: code = NotFound desc = could not find container \"fc1e3435e58e1270e767da8cf63b8cbdd4e1521ade2f9f25569aeffab472b53f\": container with ID starting with fc1e3435e58e1270e767da8cf63b8cbdd4e1521ade2f9f25569aeffab472b53f not found: ID does not exist" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.725198 4818 scope.go:117] "RemoveContainer" containerID="9051d50c2a6c41c87399558b71bdf13cf73621ff40b92ab9e2ec55df5591e3a0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.725821 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.725981 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.726091 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.726198 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9051d50c2a6c41c87399558b71bdf13cf73621ff40b92ab9e2ec55df5591e3a0"} err="failed to get container status \"9051d50c2a6c41c87399558b71bdf13cf73621ff40b92ab9e2ec55df5591e3a0\": rpc error: code = NotFound desc = could not find container \"9051d50c2a6c41c87399558b71bdf13cf73621ff40b92ab9e2ec55df5591e3a0\": container with ID starting with 9051d50c2a6c41c87399558b71bdf13cf73621ff40b92ab9e2ec55df5591e3a0 not found: ID does not exist" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.726657 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-logs\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.726686 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw959\" (UniqueName: \"kubernetes.io/projected/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-kube-api-access-vw959\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.726703 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-config-data\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.726721 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-config-data-custom\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.726739 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-etc-machine-id\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.726753 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.726822 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-internal-tls-certs\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.726864 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-scripts\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.726882 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-public-tls-certs\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.739006 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.828152 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-logs\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.828207 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw959\" (UniqueName: \"kubernetes.io/projected/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-kube-api-access-vw959\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.828226 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-config-data\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.828244 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-config-data-custom\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.828333 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-etc-machine-id\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.828350 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.828431 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-internal-tls-certs\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.828479 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-scripts\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.828499 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-public-tls-certs\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.829488 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-logs\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.832079 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-config-data\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.832138 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-etc-machine-id\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.833840 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.833974 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-config-data-custom\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.834630 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-public-tls-certs\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.836643 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-scripts\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.842077 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-internal-tls-certs\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:06 crc kubenswrapper[4818]: I1122 05:46:06.855781 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw959\" (UniqueName: \"kubernetes.io/projected/7feb3390-f51a-43f4-86c6-f44c30ebe2ab-kube-api-access-vw959\") pod \"manila-api-0\" (UID: \"7feb3390-f51a-43f4-86c6-f44c30ebe2ab\") " pod="openstack/manila-api-0" Nov 22 05:46:07 crc kubenswrapper[4818]: I1122 05:46:07.084232 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 22 05:46:07 crc kubenswrapper[4818]: I1122 05:46:07.658851 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 22 05:46:07 crc kubenswrapper[4818]: W1122 05:46:07.683238 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7feb3390_f51a_43f4_86c6_f44c30ebe2ab.slice/crio-14acd99a2e1ca9ae4b2f1dc9f1d41227f3ad1e072a4b36a232d0d47ce553569c WatchSource:0}: Error finding container 14acd99a2e1ca9ae4b2f1dc9f1d41227f3ad1e072a4b36a232d0d47ce553569c: Status 404 returned error can't find the container with id 14acd99a2e1ca9ae4b2f1dc9f1d41227f3ad1e072a4b36a232d0d47ce553569c Nov 22 05:46:07 crc kubenswrapper[4818]: I1122 05:46:07.888588 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 22 05:46:08 crc kubenswrapper[4818]: I1122 05:46:08.072874 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 22 05:46:08 crc kubenswrapper[4818]: I1122 05:46:08.120524 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:46:08 crc kubenswrapper[4818]: I1122 05:46:08.121018 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" containerName="ceilometer-central-agent" containerID="cri-o://91ce5cbe3ab722e8bc8fdc17560fba78094af6ccf82882c2e3c70847e57ed074" gracePeriod=30 Nov 22 05:46:08 crc kubenswrapper[4818]: I1122 05:46:08.121137 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" containerName="ceilometer-notification-agent" containerID="cri-o://d8a0f61bcb26e6c51fa6223cd0bff31d90f370eddb5ed5988e7c1af901ec93b9" gracePeriod=30 Nov 22 05:46:08 crc kubenswrapper[4818]: I1122 05:46:08.121170 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" containerName="sg-core" containerID="cri-o://319960ceed8832f65e4612ecc57252b581b471b3838923554855f0a059d72e2c" gracePeriod=30 Nov 22 05:46:08 crc kubenswrapper[4818]: I1122 05:46:08.121216 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" containerName="proxy-httpd" containerID="cri-o://634cc8f12feb62b2d0d5acc75244240144520c1342018fc39725c9df8ce14ecb" gracePeriod=30 Nov 22 05:46:08 crc kubenswrapper[4818]: I1122 05:46:08.306654 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="860fd889-3c2c-4202-b383-5491ca2d4e9d" path="/var/lib/kubelet/pods/860fd889-3c2c-4202-b383-5491ca2d4e9d/volumes" Nov 22 05:46:08 crc kubenswrapper[4818]: I1122 05:46:08.677816 4818 generic.go:334] "Generic (PLEG): container finished" podID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" containerID="634cc8f12feb62b2d0d5acc75244240144520c1342018fc39725c9df8ce14ecb" exitCode=0 Nov 22 05:46:08 crc kubenswrapper[4818]: I1122 05:46:08.677854 4818 generic.go:334] "Generic (PLEG): container finished" podID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" containerID="319960ceed8832f65e4612ecc57252b581b471b3838923554855f0a059d72e2c" exitCode=2 Nov 22 05:46:08 crc kubenswrapper[4818]: I1122 05:46:08.677868 4818 generic.go:334] "Generic (PLEG): container finished" podID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" containerID="91ce5cbe3ab722e8bc8fdc17560fba78094af6ccf82882c2e3c70847e57ed074" exitCode=0 Nov 22 05:46:08 crc kubenswrapper[4818]: I1122 05:46:08.677937 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b","Type":"ContainerDied","Data":"634cc8f12feb62b2d0d5acc75244240144520c1342018fc39725c9df8ce14ecb"} Nov 22 05:46:08 crc kubenswrapper[4818]: I1122 05:46:08.677970 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b","Type":"ContainerDied","Data":"319960ceed8832f65e4612ecc57252b581b471b3838923554855f0a059d72e2c"} Nov 22 05:46:08 crc kubenswrapper[4818]: I1122 05:46:08.677983 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b","Type":"ContainerDied","Data":"91ce5cbe3ab722e8bc8fdc17560fba78094af6ccf82882c2e3c70847e57ed074"} Nov 22 05:46:08 crc kubenswrapper[4818]: I1122 05:46:08.680322 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"7feb3390-f51a-43f4-86c6-f44c30ebe2ab","Type":"ContainerStarted","Data":"6fc7097f462fb1784ef7e22d94b3a74242b764d5669db90211f80c337ced6cdf"} Nov 22 05:46:08 crc kubenswrapper[4818]: I1122 05:46:08.680366 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"7feb3390-f51a-43f4-86c6-f44c30ebe2ab","Type":"ContainerStarted","Data":"14acd99a2e1ca9ae4b2f1dc9f1d41227f3ad1e072a4b36a232d0d47ce553569c"} Nov 22 05:46:09 crc kubenswrapper[4818]: I1122 05:46:09.690765 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"7feb3390-f51a-43f4-86c6-f44c30ebe2ab","Type":"ContainerStarted","Data":"af6c7cdd7a0685dc4a493de8939f42bab56413a47d3841f730a88a2f894eafd6"} Nov 22 05:46:09 crc kubenswrapper[4818]: I1122 05:46:09.691034 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Nov 22 05:46:09 crc kubenswrapper[4818]: I1122 05:46:09.715076 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.715058396 podStartE2EDuration="3.715058396s" podCreationTimestamp="2025-11-22 05:46:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:46:09.710106805 +0000 UTC m=+3522.284523342" watchObservedRunningTime="2025-11-22 05:46:09.715058396 +0000 UTC m=+3522.289474923" Nov 22 05:46:11 crc kubenswrapper[4818]: I1122 05:46:11.033684 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Nov 22 05:46:11 crc kubenswrapper[4818]: I1122 05:46:11.319433 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76b5fdb995-zl4gj" Nov 22 05:46:11 crc kubenswrapper[4818]: I1122 05:46:11.382535 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-6xc6s"] Nov 22 05:46:11 crc kubenswrapper[4818]: I1122 05:46:11.382781 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" podUID="ac91350e-274b-422f-aada-5c4f42800ffd" containerName="dnsmasq-dns" containerID="cri-o://4a6bd71c9a869a93ca44b3b32c421e5cf41ce27e9689ad6cdf9a4e39b64dabfb" gracePeriod=10 Nov 22 05:46:11 crc kubenswrapper[4818]: I1122 05:46:11.714152 4818 generic.go:334] "Generic (PLEG): container finished" podID="ac91350e-274b-422f-aada-5c4f42800ffd" containerID="4a6bd71c9a869a93ca44b3b32c421e5cf41ce27e9689ad6cdf9a4e39b64dabfb" exitCode=0 Nov 22 05:46:11 crc kubenswrapper[4818]: I1122 05:46:11.714197 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" event={"ID":"ac91350e-274b-422f-aada-5c4f42800ffd","Type":"ContainerDied","Data":"4a6bd71c9a869a93ca44b3b32c421e5cf41ce27e9689ad6cdf9a4e39b64dabfb"} Nov 22 05:46:11 crc kubenswrapper[4818]: I1122 05:46:11.907814 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:46:11 crc kubenswrapper[4818]: I1122 05:46:11.969585 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-ovsdbserver-sb\") pod \"ac91350e-274b-422f-aada-5c4f42800ffd\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " Nov 22 05:46:11 crc kubenswrapper[4818]: I1122 05:46:11.969697 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-openstack-edpm-ipam\") pod \"ac91350e-274b-422f-aada-5c4f42800ffd\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " Nov 22 05:46:11 crc kubenswrapper[4818]: I1122 05:46:11.969735 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-config\") pod \"ac91350e-274b-422f-aada-5c4f42800ffd\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " Nov 22 05:46:11 crc kubenswrapper[4818]: I1122 05:46:11.969812 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-dns-svc\") pod \"ac91350e-274b-422f-aada-5c4f42800ffd\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " Nov 22 05:46:11 crc kubenswrapper[4818]: I1122 05:46:11.970030 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-ovsdbserver-nb\") pod \"ac91350e-274b-422f-aada-5c4f42800ffd\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " Nov 22 05:46:11 crc kubenswrapper[4818]: I1122 05:46:11.970106 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bchlt\" (UniqueName: \"kubernetes.io/projected/ac91350e-274b-422f-aada-5c4f42800ffd-kube-api-access-bchlt\") pod \"ac91350e-274b-422f-aada-5c4f42800ffd\" (UID: \"ac91350e-274b-422f-aada-5c4f42800ffd\") " Nov 22 05:46:11 crc kubenswrapper[4818]: I1122 05:46:11.974530 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac91350e-274b-422f-aada-5c4f42800ffd-kube-api-access-bchlt" (OuterVolumeSpecName: "kube-api-access-bchlt") pod "ac91350e-274b-422f-aada-5c4f42800ffd" (UID: "ac91350e-274b-422f-aada-5c4f42800ffd"). InnerVolumeSpecName "kube-api-access-bchlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:46:12 crc kubenswrapper[4818]: I1122 05:46:12.036464 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "ac91350e-274b-422f-aada-5c4f42800ffd" (UID: "ac91350e-274b-422f-aada-5c4f42800ffd"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:46:12 crc kubenswrapper[4818]: I1122 05:46:12.049915 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-config" (OuterVolumeSpecName: "config") pod "ac91350e-274b-422f-aada-5c4f42800ffd" (UID: "ac91350e-274b-422f-aada-5c4f42800ffd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:46:12 crc kubenswrapper[4818]: I1122 05:46:12.060444 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ac91350e-274b-422f-aada-5c4f42800ffd" (UID: "ac91350e-274b-422f-aada-5c4f42800ffd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:46:12 crc kubenswrapper[4818]: I1122 05:46:12.073681 4818 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:12 crc kubenswrapper[4818]: I1122 05:46:12.073713 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-config\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:12 crc kubenswrapper[4818]: I1122 05:46:12.073721 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:12 crc kubenswrapper[4818]: I1122 05:46:12.073730 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bchlt\" (UniqueName: \"kubernetes.io/projected/ac91350e-274b-422f-aada-5c4f42800ffd-kube-api-access-bchlt\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:12 crc kubenswrapper[4818]: I1122 05:46:12.076074 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ac91350e-274b-422f-aada-5c4f42800ffd" (UID: "ac91350e-274b-422f-aada-5c4f42800ffd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:46:12 crc kubenswrapper[4818]: I1122 05:46:12.077776 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ac91350e-274b-422f-aada-5c4f42800ffd" (UID: "ac91350e-274b-422f-aada-5c4f42800ffd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:46:12 crc kubenswrapper[4818]: I1122 05:46:12.175000 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:12 crc kubenswrapper[4818]: I1122 05:46:12.175024 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac91350e-274b-422f-aada-5c4f42800ffd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:12 crc kubenswrapper[4818]: I1122 05:46:12.729744 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"46554362-2c6d-45b0-b725-e32873db8d22","Type":"ContainerStarted","Data":"8073ec086aada98e30c7ecdc5dee9ad2e9cba6368dbf554e3db543698e9587fc"} Nov 22 05:46:12 crc kubenswrapper[4818]: I1122 05:46:12.730088 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"46554362-2c6d-45b0-b725-e32873db8d22","Type":"ContainerStarted","Data":"3955a5df80db77c3a24f2787390471f727a6289242ce236c6b0a17d4018fba5a"} Nov 22 05:46:12 crc kubenswrapper[4818]: I1122 05:46:12.734208 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" event={"ID":"ac91350e-274b-422f-aada-5c4f42800ffd","Type":"ContainerDied","Data":"1d90dfd004e65507b32c100c598435d8882baffb8e04b296691b45f061eeccb4"} Nov 22 05:46:12 crc kubenswrapper[4818]: I1122 05:46:12.734297 4818 scope.go:117] "RemoveContainer" containerID="4a6bd71c9a869a93ca44b3b32c421e5cf41ce27e9689ad6cdf9a4e39b64dabfb" Nov 22 05:46:12 crc kubenswrapper[4818]: I1122 05:46:12.734457 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-6xc6s" Nov 22 05:46:12 crc kubenswrapper[4818]: I1122 05:46:12.767330 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.670717739 podStartE2EDuration="12.767303781s" podCreationTimestamp="2025-11-22 05:46:00 +0000 UTC" firstStartedPulling="2025-11-22 05:46:01.999025116 +0000 UTC m=+3514.573441643" lastFinishedPulling="2025-11-22 05:46:11.095611158 +0000 UTC m=+3523.670027685" observedRunningTime="2025-11-22 05:46:12.761475676 +0000 UTC m=+3525.335892223" watchObservedRunningTime="2025-11-22 05:46:12.767303781 +0000 UTC m=+3525.341720328" Nov 22 05:46:12 crc kubenswrapper[4818]: I1122 05:46:12.798569 4818 scope.go:117] "RemoveContainer" containerID="f9e5ade1c43acf64c258edc95ee2f2bc24022eceb0dadc49922d0961b9087223" Nov 22 05:46:12 crc kubenswrapper[4818]: I1122 05:46:12.805564 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-6xc6s"] Nov 22 05:46:12 crc kubenswrapper[4818]: I1122 05:46:12.830632 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-6xc6s"] Nov 22 05:46:12 crc kubenswrapper[4818]: E1122 05:46:12.917383 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3759d20_d679_42de_9a4e_f32305ec1ca0.slice/crio-aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3759d20_d679_42de_9a4e_f32305ec1ca0.slice/crio-conmon-aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514.scope\": RecentStats: unable to find data in memory cache]" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.169889 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.332580 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-log-httpd\") pod \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.332693 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-scripts\") pod \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.332757 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cdlm\" (UniqueName: \"kubernetes.io/projected/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-kube-api-access-7cdlm\") pod \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.332788 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-sg-core-conf-yaml\") pod \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.332861 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-config-data\") pod \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.332918 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-combined-ca-bundle\") pod \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.332950 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-run-httpd\") pod \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.333103 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-ceilometer-tls-certs\") pod \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\" (UID: \"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b\") " Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.333744 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" (UID: "a4350ec7-01ae-4bcc-a5f1-08cfb54e646b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.333888 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" (UID: "a4350ec7-01ae-4bcc-a5f1-08cfb54e646b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.352337 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-scripts" (OuterVolumeSpecName: "scripts") pod "a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" (UID: "a4350ec7-01ae-4bcc-a5f1-08cfb54e646b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.355381 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-kube-api-access-7cdlm" (OuterVolumeSpecName: "kube-api-access-7cdlm") pod "a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" (UID: "a4350ec7-01ae-4bcc-a5f1-08cfb54e646b"). InnerVolumeSpecName "kube-api-access-7cdlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.365986 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" (UID: "a4350ec7-01ae-4bcc-a5f1-08cfb54e646b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.427393 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" (UID: "a4350ec7-01ae-4bcc-a5f1-08cfb54e646b"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.436693 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cdlm\" (UniqueName: \"kubernetes.io/projected/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-kube-api-access-7cdlm\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.436728 4818 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.436739 4818 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.436749 4818 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.436757 4818 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.436766 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.455395 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" (UID: "a4350ec7-01ae-4bcc-a5f1-08cfb54e646b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.479523 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-config-data" (OuterVolumeSpecName: "config-data") pod "a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" (UID: "a4350ec7-01ae-4bcc-a5f1-08cfb54e646b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.539650 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.539691 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.633087 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.645809 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.748211 4818 generic.go:334] "Generic (PLEG): container finished" podID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" containerID="d8a0f61bcb26e6c51fa6223cd0bff31d90f370eddb5ed5988e7c1af901ec93b9" exitCode=0 Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.748272 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b","Type":"ContainerDied","Data":"d8a0f61bcb26e6c51fa6223cd0bff31d90f370eddb5ed5988e7c1af901ec93b9"} Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.748323 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a4350ec7-01ae-4bcc-a5f1-08cfb54e646b","Type":"ContainerDied","Data":"9e4c64805b1f7cbcf57a08669bc330fafdbd9272783cd0780077446e44e413f5"} Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.748346 4818 scope.go:117] "RemoveContainer" containerID="634cc8f12feb62b2d0d5acc75244240144520c1342018fc39725c9df8ce14ecb" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.748410 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.803310 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.805441 4818 scope.go:117] "RemoveContainer" containerID="319960ceed8832f65e4612ecc57252b581b471b3838923554855f0a059d72e2c" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.815825 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.840322 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:46:13 crc kubenswrapper[4818]: E1122 05:46:13.840763 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" containerName="sg-core" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.840782 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" containerName="sg-core" Nov 22 05:46:13 crc kubenswrapper[4818]: E1122 05:46:13.840815 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac91350e-274b-422f-aada-5c4f42800ffd" containerName="init" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.840823 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac91350e-274b-422f-aada-5c4f42800ffd" containerName="init" Nov 22 05:46:13 crc kubenswrapper[4818]: E1122 05:46:13.840836 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" containerName="ceilometer-notification-agent" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.840844 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" containerName="ceilometer-notification-agent" Nov 22 05:46:13 crc kubenswrapper[4818]: E1122 05:46:13.840854 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" containerName="proxy-httpd" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.840860 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" containerName="proxy-httpd" Nov 22 05:46:13 crc kubenswrapper[4818]: E1122 05:46:13.840870 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" containerName="ceilometer-central-agent" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.840878 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" containerName="ceilometer-central-agent" Nov 22 05:46:13 crc kubenswrapper[4818]: E1122 05:46:13.840892 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac91350e-274b-422f-aada-5c4f42800ffd" containerName="dnsmasq-dns" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.840898 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac91350e-274b-422f-aada-5c4f42800ffd" containerName="dnsmasq-dns" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.841093 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" containerName="proxy-httpd" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.841113 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" containerName="ceilometer-central-agent" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.841133 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" containerName="ceilometer-notification-agent" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.841151 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac91350e-274b-422f-aada-5c4f42800ffd" containerName="dnsmasq-dns" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.841166 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" containerName="sg-core" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.843445 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.850118 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.852028 4818 scope.go:117] "RemoveContainer" containerID="d8a0f61bcb26e6c51fa6223cd0bff31d90f370eddb5ed5988e7c1af901ec93b9" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.864062 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.867456 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.867655 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.882612 4818 scope.go:117] "RemoveContainer" containerID="91ce5cbe3ab722e8bc8fdc17560fba78094af6ccf82882c2e3c70847e57ed074" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.910323 4818 scope.go:117] "RemoveContainer" containerID="634cc8f12feb62b2d0d5acc75244240144520c1342018fc39725c9df8ce14ecb" Nov 22 05:46:13 crc kubenswrapper[4818]: E1122 05:46:13.911769 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"634cc8f12feb62b2d0d5acc75244240144520c1342018fc39725c9df8ce14ecb\": container with ID starting with 634cc8f12feb62b2d0d5acc75244240144520c1342018fc39725c9df8ce14ecb not found: ID does not exist" containerID="634cc8f12feb62b2d0d5acc75244240144520c1342018fc39725c9df8ce14ecb" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.911824 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"634cc8f12feb62b2d0d5acc75244240144520c1342018fc39725c9df8ce14ecb"} err="failed to get container status \"634cc8f12feb62b2d0d5acc75244240144520c1342018fc39725c9df8ce14ecb\": rpc error: code = NotFound desc = could not find container \"634cc8f12feb62b2d0d5acc75244240144520c1342018fc39725c9df8ce14ecb\": container with ID starting with 634cc8f12feb62b2d0d5acc75244240144520c1342018fc39725c9df8ce14ecb not found: ID does not exist" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.911855 4818 scope.go:117] "RemoveContainer" containerID="319960ceed8832f65e4612ecc57252b581b471b3838923554855f0a059d72e2c" Nov 22 05:46:13 crc kubenswrapper[4818]: E1122 05:46:13.912121 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"319960ceed8832f65e4612ecc57252b581b471b3838923554855f0a059d72e2c\": container with ID starting with 319960ceed8832f65e4612ecc57252b581b471b3838923554855f0a059d72e2c not found: ID does not exist" containerID="319960ceed8832f65e4612ecc57252b581b471b3838923554855f0a059d72e2c" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.912151 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"319960ceed8832f65e4612ecc57252b581b471b3838923554855f0a059d72e2c"} err="failed to get container status \"319960ceed8832f65e4612ecc57252b581b471b3838923554855f0a059d72e2c\": rpc error: code = NotFound desc = could not find container \"319960ceed8832f65e4612ecc57252b581b471b3838923554855f0a059d72e2c\": container with ID starting with 319960ceed8832f65e4612ecc57252b581b471b3838923554855f0a059d72e2c not found: ID does not exist" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.912172 4818 scope.go:117] "RemoveContainer" containerID="d8a0f61bcb26e6c51fa6223cd0bff31d90f370eddb5ed5988e7c1af901ec93b9" Nov 22 05:46:13 crc kubenswrapper[4818]: E1122 05:46:13.912473 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8a0f61bcb26e6c51fa6223cd0bff31d90f370eddb5ed5988e7c1af901ec93b9\": container with ID starting with d8a0f61bcb26e6c51fa6223cd0bff31d90f370eddb5ed5988e7c1af901ec93b9 not found: ID does not exist" containerID="d8a0f61bcb26e6c51fa6223cd0bff31d90f370eddb5ed5988e7c1af901ec93b9" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.912497 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8a0f61bcb26e6c51fa6223cd0bff31d90f370eddb5ed5988e7c1af901ec93b9"} err="failed to get container status \"d8a0f61bcb26e6c51fa6223cd0bff31d90f370eddb5ed5988e7c1af901ec93b9\": rpc error: code = NotFound desc = could not find container \"d8a0f61bcb26e6c51fa6223cd0bff31d90f370eddb5ed5988e7c1af901ec93b9\": container with ID starting with d8a0f61bcb26e6c51fa6223cd0bff31d90f370eddb5ed5988e7c1af901ec93b9 not found: ID does not exist" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.912510 4818 scope.go:117] "RemoveContainer" containerID="91ce5cbe3ab722e8bc8fdc17560fba78094af6ccf82882c2e3c70847e57ed074" Nov 22 05:46:13 crc kubenswrapper[4818]: E1122 05:46:13.912673 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91ce5cbe3ab722e8bc8fdc17560fba78094af6ccf82882c2e3c70847e57ed074\": container with ID starting with 91ce5cbe3ab722e8bc8fdc17560fba78094af6ccf82882c2e3c70847e57ed074 not found: ID does not exist" containerID="91ce5cbe3ab722e8bc8fdc17560fba78094af6ccf82882c2e3c70847e57ed074" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.912691 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91ce5cbe3ab722e8bc8fdc17560fba78094af6ccf82882c2e3c70847e57ed074"} err="failed to get container status \"91ce5cbe3ab722e8bc8fdc17560fba78094af6ccf82882c2e3c70847e57ed074\": rpc error: code = NotFound desc = could not find container \"91ce5cbe3ab722e8bc8fdc17560fba78094af6ccf82882c2e3c70847e57ed074\": container with ID starting with 91ce5cbe3ab722e8bc8fdc17560fba78094af6ccf82882c2e3c70847e57ed074 not found: ID does not exist" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.946811 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.946871 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-scripts\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.946898 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xscld\" (UniqueName: \"kubernetes.io/projected/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-kube-api-access-xscld\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.946936 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.947006 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-log-httpd\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.947052 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.947079 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-config-data\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:13 crc kubenswrapper[4818]: I1122 05:46:13.947100 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-run-httpd\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.049134 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-log-httpd\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.049197 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.049233 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-config-data\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.049276 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-run-httpd\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.049325 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.049733 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-run-httpd\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.049793 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-scripts\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.049796 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-log-httpd\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.049849 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xscld\" (UniqueName: \"kubernetes.io/projected/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-kube-api-access-xscld\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.049903 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.055026 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.055049 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.055281 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-config-data\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.055559 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-scripts\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.056410 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.069847 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xscld\" (UniqueName: \"kubernetes.io/projected/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-kube-api-access-xscld\") pod \"ceilometer-0\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " pod="openstack/ceilometer-0" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.173450 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.347911 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4350ec7-01ae-4bcc-a5f1-08cfb54e646b" path="/var/lib/kubelet/pods/a4350ec7-01ae-4bcc-a5f1-08cfb54e646b/volumes" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.348999 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac91350e-274b-422f-aada-5c4f42800ffd" path="/var/lib/kubelet/pods/ac91350e-274b-422f-aada-5c4f42800ffd/volumes" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.429469 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.670598 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f784b6cfc-t55xf" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.741767 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.769437 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c4458788c-rwwjv" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.772036 4818 generic.go:334] "Generic (PLEG): container finished" podID="d03bb0b9-32cd-43b8-92e4-2b8f83361246" containerID="fb8ef4485cc8560984370390d18efaff10c06a33c0b320e64cb2d496a0f8109b" exitCode=137 Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.772068 4818 generic.go:334] "Generic (PLEG): container finished" podID="d03bb0b9-32cd-43b8-92e4-2b8f83361246" containerID="03e2148476f595fa4c74ff9412cdb8bfb41d861a81114fa9b7642fafe4b21c70" exitCode=137 Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.772105 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f784b6cfc-t55xf" event={"ID":"d03bb0b9-32cd-43b8-92e4-2b8f83361246","Type":"ContainerDied","Data":"fb8ef4485cc8560984370390d18efaff10c06a33c0b320e64cb2d496a0f8109b"} Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.772131 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f784b6cfc-t55xf" event={"ID":"d03bb0b9-32cd-43b8-92e4-2b8f83361246","Type":"ContainerDied","Data":"03e2148476f595fa4c74ff9412cdb8bfb41d861a81114fa9b7642fafe4b21c70"} Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.772141 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f784b6cfc-t55xf" event={"ID":"d03bb0b9-32cd-43b8-92e4-2b8f83361246","Type":"ContainerDied","Data":"46ebc9a94338c8ec0c1da411417b3b80dc639867a75e00d379558ce3aaeedc22"} Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.772155 4818 scope.go:117] "RemoveContainer" containerID="fb8ef4485cc8560984370390d18efaff10c06a33c0b320e64cb2d496a0f8109b" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.772297 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f784b6cfc-t55xf" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.776422 4818 generic.go:334] "Generic (PLEG): container finished" podID="90e9c626-fc6c-4ad9-b385-e634bd3dbfb8" containerID="39f6a96f45c4ddd4908190f409436048cbd94f00d707a0d23d84995b34baa4ef" exitCode=137 Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.776453 4818 generic.go:334] "Generic (PLEG): container finished" podID="90e9c626-fc6c-4ad9-b385-e634bd3dbfb8" containerID="c169439386f8d7a558f2d85a41494b3e0548e5681b0841253f085f632029c464" exitCode=137 Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.776542 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c4458788c-rwwjv" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.777058 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c4458788c-rwwjv" event={"ID":"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8","Type":"ContainerDied","Data":"39f6a96f45c4ddd4908190f409436048cbd94f00d707a0d23d84995b34baa4ef"} Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.777088 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c4458788c-rwwjv" event={"ID":"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8","Type":"ContainerDied","Data":"c169439386f8d7a558f2d85a41494b3e0548e5681b0841253f085f632029c464"} Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.779965 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4dd89d48-7c16-45d5-9b2c-490fd14d2a74","Type":"ContainerStarted","Data":"21344755565f3718753124bf1ed45c7ff13bfa5c83b567f1cce51c996b27d9a8"} Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.868228 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d03bb0b9-32cd-43b8-92e4-2b8f83361246-config-data\") pod \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\" (UID: \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\") " Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.868291 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dh594\" (UniqueName: \"kubernetes.io/projected/d03bb0b9-32cd-43b8-92e4-2b8f83361246-kube-api-access-dh594\") pod \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\" (UID: \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\") " Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.868430 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-horizon-secret-key\") pod \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\" (UID: \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\") " Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.868476 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d03bb0b9-32cd-43b8-92e4-2b8f83361246-scripts\") pod \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\" (UID: \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\") " Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.868513 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-logs\") pod \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\" (UID: \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\") " Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.868545 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d03bb0b9-32cd-43b8-92e4-2b8f83361246-logs\") pod \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\" (UID: \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\") " Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.868596 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-config-data\") pod \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\" (UID: \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\") " Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.868652 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ks98r\" (UniqueName: \"kubernetes.io/projected/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-kube-api-access-ks98r\") pod \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\" (UID: \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\") " Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.868690 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d03bb0b9-32cd-43b8-92e4-2b8f83361246-horizon-secret-key\") pod \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\" (UID: \"d03bb0b9-32cd-43b8-92e4-2b8f83361246\") " Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.868715 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-scripts\") pod \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\" (UID: \"90e9c626-fc6c-4ad9-b385-e634bd3dbfb8\") " Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.869431 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-logs" (OuterVolumeSpecName: "logs") pod "90e9c626-fc6c-4ad9-b385-e634bd3dbfb8" (UID: "90e9c626-fc6c-4ad9-b385-e634bd3dbfb8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.870873 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d03bb0b9-32cd-43b8-92e4-2b8f83361246-logs" (OuterVolumeSpecName: "logs") pod "d03bb0b9-32cd-43b8-92e4-2b8f83361246" (UID: "d03bb0b9-32cd-43b8-92e4-2b8f83361246"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.876909 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d03bb0b9-32cd-43b8-92e4-2b8f83361246-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d03bb0b9-32cd-43b8-92e4-2b8f83361246" (UID: "d03bb0b9-32cd-43b8-92e4-2b8f83361246"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.879695 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d03bb0b9-32cd-43b8-92e4-2b8f83361246-kube-api-access-dh594" (OuterVolumeSpecName: "kube-api-access-dh594") pod "d03bb0b9-32cd-43b8-92e4-2b8f83361246" (UID: "d03bb0b9-32cd-43b8-92e4-2b8f83361246"). InnerVolumeSpecName "kube-api-access-dh594". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.880333 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "90e9c626-fc6c-4ad9-b385-e634bd3dbfb8" (UID: "90e9c626-fc6c-4ad9-b385-e634bd3dbfb8"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.883531 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-kube-api-access-ks98r" (OuterVolumeSpecName: "kube-api-access-ks98r") pod "90e9c626-fc6c-4ad9-b385-e634bd3dbfb8" (UID: "90e9c626-fc6c-4ad9-b385-e634bd3dbfb8"). InnerVolumeSpecName "kube-api-access-ks98r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.902624 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d03bb0b9-32cd-43b8-92e4-2b8f83361246-config-data" (OuterVolumeSpecName: "config-data") pod "d03bb0b9-32cd-43b8-92e4-2b8f83361246" (UID: "d03bb0b9-32cd-43b8-92e4-2b8f83361246"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.903238 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-config-data" (OuterVolumeSpecName: "config-data") pod "90e9c626-fc6c-4ad9-b385-e634bd3dbfb8" (UID: "90e9c626-fc6c-4ad9-b385-e634bd3dbfb8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.913551 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-scripts" (OuterVolumeSpecName: "scripts") pod "90e9c626-fc6c-4ad9-b385-e634bd3dbfb8" (UID: "90e9c626-fc6c-4ad9-b385-e634bd3dbfb8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.915825 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d03bb0b9-32cd-43b8-92e4-2b8f83361246-scripts" (OuterVolumeSpecName: "scripts") pod "d03bb0b9-32cd-43b8-92e4-2b8f83361246" (UID: "d03bb0b9-32cd-43b8-92e4-2b8f83361246"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.970618 4818 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d03bb0b9-32cd-43b8-92e4-2b8f83361246-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.970666 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.970678 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d03bb0b9-32cd-43b8-92e4-2b8f83361246-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.970690 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dh594\" (UniqueName: \"kubernetes.io/projected/d03bb0b9-32cd-43b8-92e4-2b8f83361246-kube-api-access-dh594\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.970704 4818 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.970715 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d03bb0b9-32cd-43b8-92e4-2b8f83361246-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.970725 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-logs\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.970735 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d03bb0b9-32cd-43b8-92e4-2b8f83361246-logs\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.970750 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.970762 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ks98r\" (UniqueName: \"kubernetes.io/projected/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8-kube-api-access-ks98r\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:14 crc kubenswrapper[4818]: I1122 05:46:14.978337 4818 scope.go:117] "RemoveContainer" containerID="03e2148476f595fa4c74ff9412cdb8bfb41d861a81114fa9b7642fafe4b21c70" Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.008882 4818 scope.go:117] "RemoveContainer" containerID="fb8ef4485cc8560984370390d18efaff10c06a33c0b320e64cb2d496a0f8109b" Nov 22 05:46:15 crc kubenswrapper[4818]: E1122 05:46:15.009677 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb8ef4485cc8560984370390d18efaff10c06a33c0b320e64cb2d496a0f8109b\": container with ID starting with fb8ef4485cc8560984370390d18efaff10c06a33c0b320e64cb2d496a0f8109b not found: ID does not exist" containerID="fb8ef4485cc8560984370390d18efaff10c06a33c0b320e64cb2d496a0f8109b" Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.009730 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb8ef4485cc8560984370390d18efaff10c06a33c0b320e64cb2d496a0f8109b"} err="failed to get container status \"fb8ef4485cc8560984370390d18efaff10c06a33c0b320e64cb2d496a0f8109b\": rpc error: code = NotFound desc = could not find container \"fb8ef4485cc8560984370390d18efaff10c06a33c0b320e64cb2d496a0f8109b\": container with ID starting with fb8ef4485cc8560984370390d18efaff10c06a33c0b320e64cb2d496a0f8109b not found: ID does not exist" Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.009755 4818 scope.go:117] "RemoveContainer" containerID="03e2148476f595fa4c74ff9412cdb8bfb41d861a81114fa9b7642fafe4b21c70" Nov 22 05:46:15 crc kubenswrapper[4818]: E1122 05:46:15.010416 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03e2148476f595fa4c74ff9412cdb8bfb41d861a81114fa9b7642fafe4b21c70\": container with ID starting with 03e2148476f595fa4c74ff9412cdb8bfb41d861a81114fa9b7642fafe4b21c70 not found: ID does not exist" containerID="03e2148476f595fa4c74ff9412cdb8bfb41d861a81114fa9b7642fafe4b21c70" Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.010453 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03e2148476f595fa4c74ff9412cdb8bfb41d861a81114fa9b7642fafe4b21c70"} err="failed to get container status \"03e2148476f595fa4c74ff9412cdb8bfb41d861a81114fa9b7642fafe4b21c70\": rpc error: code = NotFound desc = could not find container \"03e2148476f595fa4c74ff9412cdb8bfb41d861a81114fa9b7642fafe4b21c70\": container with ID starting with 03e2148476f595fa4c74ff9412cdb8bfb41d861a81114fa9b7642fafe4b21c70 not found: ID does not exist" Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.010480 4818 scope.go:117] "RemoveContainer" containerID="fb8ef4485cc8560984370390d18efaff10c06a33c0b320e64cb2d496a0f8109b" Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.010934 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb8ef4485cc8560984370390d18efaff10c06a33c0b320e64cb2d496a0f8109b"} err="failed to get container status \"fb8ef4485cc8560984370390d18efaff10c06a33c0b320e64cb2d496a0f8109b\": rpc error: code = NotFound desc = could not find container \"fb8ef4485cc8560984370390d18efaff10c06a33c0b320e64cb2d496a0f8109b\": container with ID starting with fb8ef4485cc8560984370390d18efaff10c06a33c0b320e64cb2d496a0f8109b not found: ID does not exist" Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.010955 4818 scope.go:117] "RemoveContainer" containerID="03e2148476f595fa4c74ff9412cdb8bfb41d861a81114fa9b7642fafe4b21c70" Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.011426 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03e2148476f595fa4c74ff9412cdb8bfb41d861a81114fa9b7642fafe4b21c70"} err="failed to get container status \"03e2148476f595fa4c74ff9412cdb8bfb41d861a81114fa9b7642fafe4b21c70\": rpc error: code = NotFound desc = could not find container \"03e2148476f595fa4c74ff9412cdb8bfb41d861a81114fa9b7642fafe4b21c70\": container with ID starting with 03e2148476f595fa4c74ff9412cdb8bfb41d861a81114fa9b7642fafe4b21c70 not found: ID does not exist" Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.011449 4818 scope.go:117] "RemoveContainer" containerID="39f6a96f45c4ddd4908190f409436048cbd94f00d707a0d23d84995b34baa4ef" Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.126023 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-f784b6cfc-t55xf"] Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.141499 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-f784b6cfc-t55xf"] Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.160407 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c4458788c-rwwjv"] Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.197504 4818 scope.go:117] "RemoveContainer" containerID="c169439386f8d7a558f2d85a41494b3e0548e5681b0841253f085f632029c464" Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.204799 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-c4458788c-rwwjv"] Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.236414 4818 scope.go:117] "RemoveContainer" containerID="39f6a96f45c4ddd4908190f409436048cbd94f00d707a0d23d84995b34baa4ef" Nov 22 05:46:15 crc kubenswrapper[4818]: E1122 05:46:15.236793 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39f6a96f45c4ddd4908190f409436048cbd94f00d707a0d23d84995b34baa4ef\": container with ID starting with 39f6a96f45c4ddd4908190f409436048cbd94f00d707a0d23d84995b34baa4ef not found: ID does not exist" containerID="39f6a96f45c4ddd4908190f409436048cbd94f00d707a0d23d84995b34baa4ef" Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.236823 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39f6a96f45c4ddd4908190f409436048cbd94f00d707a0d23d84995b34baa4ef"} err="failed to get container status \"39f6a96f45c4ddd4908190f409436048cbd94f00d707a0d23d84995b34baa4ef\": rpc error: code = NotFound desc = could not find container \"39f6a96f45c4ddd4908190f409436048cbd94f00d707a0d23d84995b34baa4ef\": container with ID starting with 39f6a96f45c4ddd4908190f409436048cbd94f00d707a0d23d84995b34baa4ef not found: ID does not exist" Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.236844 4818 scope.go:117] "RemoveContainer" containerID="c169439386f8d7a558f2d85a41494b3e0548e5681b0841253f085f632029c464" Nov 22 05:46:15 crc kubenswrapper[4818]: E1122 05:46:15.237303 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c169439386f8d7a558f2d85a41494b3e0548e5681b0841253f085f632029c464\": container with ID starting with c169439386f8d7a558f2d85a41494b3e0548e5681b0841253f085f632029c464 not found: ID does not exist" containerID="c169439386f8d7a558f2d85a41494b3e0548e5681b0841253f085f632029c464" Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.237333 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c169439386f8d7a558f2d85a41494b3e0548e5681b0841253f085f632029c464"} err="failed to get container status \"c169439386f8d7a558f2d85a41494b3e0548e5681b0841253f085f632029c464\": rpc error: code = NotFound desc = could not find container \"c169439386f8d7a558f2d85a41494b3e0548e5681b0841253f085f632029c464\": container with ID starting with c169439386f8d7a558f2d85a41494b3e0548e5681b0841253f085f632029c464 not found: ID does not exist" Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.576635 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.603853 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-86757bfcb-s9bbk" Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.684582 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7764859bb4-tnzkk"] Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.793498 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7764859bb4-tnzkk" podUID="60b9da36-1281-4cc5-a9ac-a2df2e4d9638" containerName="horizon-log" containerID="cri-o://180e7581aa51496380128694e8a8eda047b65550ed4a7f9e238af15906922636" gracePeriod=30 Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.793752 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4dd89d48-7c16-45d5-9b2c-490fd14d2a74","Type":"ContainerStarted","Data":"443e04113ac1bd8191d0c04b1e5d053e6b05e4d393f21aa7c5c1d661cae0b1d4"} Nov 22 05:46:15 crc kubenswrapper[4818]: I1122 05:46:15.793978 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7764859bb4-tnzkk" podUID="60b9da36-1281-4cc5-a9ac-a2df2e4d9638" containerName="horizon" containerID="cri-o://64ff53e988010bba9bf54bb360558aabbc0e90386763b14c6ab5fe8f007de1e8" gracePeriod=30 Nov 22 05:46:16 crc kubenswrapper[4818]: I1122 05:46:16.306015 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90e9c626-fc6c-4ad9-b385-e634bd3dbfb8" path="/var/lib/kubelet/pods/90e9c626-fc6c-4ad9-b385-e634bd3dbfb8/volumes" Nov 22 05:46:16 crc kubenswrapper[4818]: I1122 05:46:16.308591 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d03bb0b9-32cd-43b8-92e4-2b8f83361246" path="/var/lib/kubelet/pods/d03bb0b9-32cd-43b8-92e4-2b8f83361246/volumes" Nov 22 05:46:16 crc kubenswrapper[4818]: I1122 05:46:16.806660 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4dd89d48-7c16-45d5-9b2c-490fd14d2a74","Type":"ContainerStarted","Data":"a322b78ab74f284bf3442fad2ac569e3ece37f09b972e5c0d5cdfe47185c82b0"} Nov 22 05:46:16 crc kubenswrapper[4818]: I1122 05:46:16.807134 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4dd89d48-7c16-45d5-9b2c-490fd14d2a74","Type":"ContainerStarted","Data":"9d2a37c5014ff64bc4673b4171d2a276ee78a0561ed28fe20196b1fb48a9586c"} Nov 22 05:46:18 crc kubenswrapper[4818]: I1122 05:46:18.840790 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4dd89d48-7c16-45d5-9b2c-490fd14d2a74","Type":"ContainerStarted","Data":"2b77f0a8bbc806215e38b629df2378a75fd23abc7352b7d496609d2052ccd5a2"} Nov 22 05:46:18 crc kubenswrapper[4818]: I1122 05:46:18.841519 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 05:46:18 crc kubenswrapper[4818]: I1122 05:46:18.841234 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" containerName="proxy-httpd" containerID="cri-o://2b77f0a8bbc806215e38b629df2378a75fd23abc7352b7d496609d2052ccd5a2" gracePeriod=30 Nov 22 05:46:18 crc kubenswrapper[4818]: I1122 05:46:18.841206 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" containerName="sg-core" containerID="cri-o://a322b78ab74f284bf3442fad2ac569e3ece37f09b972e5c0d5cdfe47185c82b0" gracePeriod=30 Nov 22 05:46:18 crc kubenswrapper[4818]: I1122 05:46:18.841322 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" containerName="ceilometer-notification-agent" containerID="cri-o://9d2a37c5014ff64bc4673b4171d2a276ee78a0561ed28fe20196b1fb48a9586c" gracePeriod=30 Nov 22 05:46:18 crc kubenswrapper[4818]: I1122 05:46:18.841891 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" containerName="ceilometer-central-agent" containerID="cri-o://443e04113ac1bd8191d0c04b1e5d053e6b05e4d393f21aa7c5c1d661cae0b1d4" gracePeriod=30 Nov 22 05:46:19 crc kubenswrapper[4818]: I1122 05:46:19.858862 4818 generic.go:334] "Generic (PLEG): container finished" podID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" containerID="2b77f0a8bbc806215e38b629df2378a75fd23abc7352b7d496609d2052ccd5a2" exitCode=0 Nov 22 05:46:19 crc kubenswrapper[4818]: I1122 05:46:19.858916 4818 generic.go:334] "Generic (PLEG): container finished" podID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" containerID="a322b78ab74f284bf3442fad2ac569e3ece37f09b972e5c0d5cdfe47185c82b0" exitCode=2 Nov 22 05:46:19 crc kubenswrapper[4818]: I1122 05:46:19.858940 4818 generic.go:334] "Generic (PLEG): container finished" podID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" containerID="9d2a37c5014ff64bc4673b4171d2a276ee78a0561ed28fe20196b1fb48a9586c" exitCode=0 Nov 22 05:46:19 crc kubenswrapper[4818]: I1122 05:46:19.858994 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4dd89d48-7c16-45d5-9b2c-490fd14d2a74","Type":"ContainerDied","Data":"2b77f0a8bbc806215e38b629df2378a75fd23abc7352b7d496609d2052ccd5a2"} Nov 22 05:46:19 crc kubenswrapper[4818]: I1122 05:46:19.859063 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4dd89d48-7c16-45d5-9b2c-490fd14d2a74","Type":"ContainerDied","Data":"a322b78ab74f284bf3442fad2ac569e3ece37f09b972e5c0d5cdfe47185c82b0"} Nov 22 05:46:19 crc kubenswrapper[4818]: I1122 05:46:19.859084 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4dd89d48-7c16-45d5-9b2c-490fd14d2a74","Type":"ContainerDied","Data":"9d2a37c5014ff64bc4673b4171d2a276ee78a0561ed28fe20196b1fb48a9586c"} Nov 22 05:46:19 crc kubenswrapper[4818]: I1122 05:46:19.862699 4818 generic.go:334] "Generic (PLEG): container finished" podID="60b9da36-1281-4cc5-a9ac-a2df2e4d9638" containerID="64ff53e988010bba9bf54bb360558aabbc0e90386763b14c6ab5fe8f007de1e8" exitCode=0 Nov 22 05:46:19 crc kubenswrapper[4818]: I1122 05:46:19.862760 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7764859bb4-tnzkk" event={"ID":"60b9da36-1281-4cc5-a9ac-a2df2e4d9638","Type":"ContainerDied","Data":"64ff53e988010bba9bf54bb360558aabbc0e90386763b14c6ab5fe8f007de1e8"} Nov 22 05:46:20 crc kubenswrapper[4818]: I1122 05:46:20.879549 4818 generic.go:334] "Generic (PLEG): container finished" podID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" containerID="443e04113ac1bd8191d0c04b1e5d053e6b05e4d393f21aa7c5c1d661cae0b1d4" exitCode=0 Nov 22 05:46:20 crc kubenswrapper[4818]: I1122 05:46:20.879653 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4dd89d48-7c16-45d5-9b2c-490fd14d2a74","Type":"ContainerDied","Data":"443e04113ac1bd8191d0c04b1e5d053e6b05e4d393f21aa7c5c1d661cae0b1d4"} Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.102174 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7764859bb4-tnzkk" podUID="60b9da36-1281-4cc5-a9ac-a2df2e4d9638" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.242:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.242:8443: connect: connection refused" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.118576 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.210232 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.264534 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.264612 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.308885 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-scripts\") pod \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.309026 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-combined-ca-bundle\") pod \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.309092 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-run-httpd\") pod \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.309154 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-sg-core-conf-yaml\") pod \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.309198 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xscld\" (UniqueName: \"kubernetes.io/projected/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-kube-api-access-xscld\") pod \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.309228 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-log-httpd\") pod \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.309273 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-config-data\") pod \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.309318 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-ceilometer-tls-certs\") pod \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\" (UID: \"4dd89d48-7c16-45d5-9b2c-490fd14d2a74\") " Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.309603 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4dd89d48-7c16-45d5-9b2c-490fd14d2a74" (UID: "4dd89d48-7c16-45d5-9b2c-490fd14d2a74"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.310090 4818 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.310569 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4dd89d48-7c16-45d5-9b2c-490fd14d2a74" (UID: "4dd89d48-7c16-45d5-9b2c-490fd14d2a74"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.316322 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-kube-api-access-xscld" (OuterVolumeSpecName: "kube-api-access-xscld") pod "4dd89d48-7c16-45d5-9b2c-490fd14d2a74" (UID: "4dd89d48-7c16-45d5-9b2c-490fd14d2a74"). InnerVolumeSpecName "kube-api-access-xscld". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.325773 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-scripts" (OuterVolumeSpecName: "scripts") pod "4dd89d48-7c16-45d5-9b2c-490fd14d2a74" (UID: "4dd89d48-7c16-45d5-9b2c-490fd14d2a74"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.344712 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4dd89d48-7c16-45d5-9b2c-490fd14d2a74" (UID: "4dd89d48-7c16-45d5-9b2c-490fd14d2a74"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.394333 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "4dd89d48-7c16-45d5-9b2c-490fd14d2a74" (UID: "4dd89d48-7c16-45d5-9b2c-490fd14d2a74"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.414011 4818 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.414066 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xscld\" (UniqueName: \"kubernetes.io/projected/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-kube-api-access-xscld\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.414087 4818 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.414104 4818 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.414144 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.430350 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4dd89d48-7c16-45d5-9b2c-490fd14d2a74" (UID: "4dd89d48-7c16-45d5-9b2c-490fd14d2a74"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.443555 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-config-data" (OuterVolumeSpecName: "config-data") pod "4dd89d48-7c16-45d5-9b2c-490fd14d2a74" (UID: "4dd89d48-7c16-45d5-9b2c-490fd14d2a74"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.516184 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.516214 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dd89d48-7c16-45d5-9b2c-490fd14d2a74-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.893177 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4dd89d48-7c16-45d5-9b2c-490fd14d2a74","Type":"ContainerDied","Data":"21344755565f3718753124bf1ed45c7ff13bfa5c83b567f1cce51c996b27d9a8"} Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.893558 4818 scope.go:117] "RemoveContainer" containerID="2b77f0a8bbc806215e38b629df2378a75fd23abc7352b7d496609d2052ccd5a2" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.893238 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.969806 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:46:21 crc kubenswrapper[4818]: I1122 05:46:21.991196 4818 scope.go:117] "RemoveContainer" containerID="a322b78ab74f284bf3442fad2ac569e3ece37f09b972e5c0d5cdfe47185c82b0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.001111 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.020604 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:46:22 crc kubenswrapper[4818]: E1122 05:46:22.020930 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" containerName="ceilometer-central-agent" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.020941 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" containerName="ceilometer-central-agent" Nov 22 05:46:22 crc kubenswrapper[4818]: E1122 05:46:22.020952 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d03bb0b9-32cd-43b8-92e4-2b8f83361246" containerName="horizon-log" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.020959 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="d03bb0b9-32cd-43b8-92e4-2b8f83361246" containerName="horizon-log" Nov 22 05:46:22 crc kubenswrapper[4818]: E1122 05:46:22.020984 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" containerName="ceilometer-notification-agent" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.020990 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" containerName="ceilometer-notification-agent" Nov 22 05:46:22 crc kubenswrapper[4818]: E1122 05:46:22.020999 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90e9c626-fc6c-4ad9-b385-e634bd3dbfb8" containerName="horizon" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.021004 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="90e9c626-fc6c-4ad9-b385-e634bd3dbfb8" containerName="horizon" Nov 22 05:46:22 crc kubenswrapper[4818]: E1122 05:46:22.021015 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d03bb0b9-32cd-43b8-92e4-2b8f83361246" containerName="horizon" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.021021 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="d03bb0b9-32cd-43b8-92e4-2b8f83361246" containerName="horizon" Nov 22 05:46:22 crc kubenswrapper[4818]: E1122 05:46:22.021032 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90e9c626-fc6c-4ad9-b385-e634bd3dbfb8" containerName="horizon-log" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.021038 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="90e9c626-fc6c-4ad9-b385-e634bd3dbfb8" containerName="horizon-log" Nov 22 05:46:22 crc kubenswrapper[4818]: E1122 05:46:22.021048 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" containerName="proxy-httpd" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.021054 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" containerName="proxy-httpd" Nov 22 05:46:22 crc kubenswrapper[4818]: E1122 05:46:22.021065 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" containerName="sg-core" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.021071 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" containerName="sg-core" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.021223 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" containerName="proxy-httpd" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.021238 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="d03bb0b9-32cd-43b8-92e4-2b8f83361246" containerName="horizon-log" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.021269 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="d03bb0b9-32cd-43b8-92e4-2b8f83361246" containerName="horizon" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.021284 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="90e9c626-fc6c-4ad9-b385-e634bd3dbfb8" containerName="horizon-log" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.021297 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" containerName="sg-core" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.021309 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" containerName="ceilometer-central-agent" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.021319 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" containerName="ceilometer-notification-agent" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.021329 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="90e9c626-fc6c-4ad9-b385-e634bd3dbfb8" containerName="horizon" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.023031 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.027725 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.027923 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.028311 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.038346 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.054780 4818 scope.go:117] "RemoveContainer" containerID="9d2a37c5014ff64bc4673b4171d2a276ee78a0561ed28fe20196b1fb48a9586c" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.076725 4818 scope.go:117] "RemoveContainer" containerID="443e04113ac1bd8191d0c04b1e5d053e6b05e4d393f21aa7c5c1d661cae0b1d4" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.128443 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfpph\" (UniqueName: \"kubernetes.io/projected/4ea3ac76-3418-4b1c-8800-d57f554e50e3-kube-api-access-vfpph\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.128488 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ea3ac76-3418-4b1c-8800-d57f554e50e3-scripts\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.128512 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ea3ac76-3418-4b1c-8800-d57f554e50e3-config-data\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.128789 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ea3ac76-3418-4b1c-8800-d57f554e50e3-log-httpd\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.128879 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ea3ac76-3418-4b1c-8800-d57f554e50e3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.129030 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ea3ac76-3418-4b1c-8800-d57f554e50e3-run-httpd\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.129098 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ea3ac76-3418-4b1c-8800-d57f554e50e3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.129156 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ea3ac76-3418-4b1c-8800-d57f554e50e3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.231602 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ea3ac76-3418-4b1c-8800-d57f554e50e3-log-httpd\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.231691 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ea3ac76-3418-4b1c-8800-d57f554e50e3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.231782 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ea3ac76-3418-4b1c-8800-d57f554e50e3-run-httpd\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.231839 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ea3ac76-3418-4b1c-8800-d57f554e50e3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.231883 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ea3ac76-3418-4b1c-8800-d57f554e50e3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.231976 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfpph\" (UniqueName: \"kubernetes.io/projected/4ea3ac76-3418-4b1c-8800-d57f554e50e3-kube-api-access-vfpph\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.232023 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ea3ac76-3418-4b1c-8800-d57f554e50e3-scripts\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.232064 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ea3ac76-3418-4b1c-8800-d57f554e50e3-config-data\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.232173 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ea3ac76-3418-4b1c-8800-d57f554e50e3-log-httpd\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.234086 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ea3ac76-3418-4b1c-8800-d57f554e50e3-run-httpd\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.239393 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ea3ac76-3418-4b1c-8800-d57f554e50e3-scripts\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.240492 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ea3ac76-3418-4b1c-8800-d57f554e50e3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.241769 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ea3ac76-3418-4b1c-8800-d57f554e50e3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.242980 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ea3ac76-3418-4b1c-8800-d57f554e50e3-config-data\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.249732 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfpph\" (UniqueName: \"kubernetes.io/projected/4ea3ac76-3418-4b1c-8800-d57f554e50e3-kube-api-access-vfpph\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.256218 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ea3ac76-3418-4b1c-8800-d57f554e50e3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4ea3ac76-3418-4b1c-8800-d57f554e50e3\") " pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.301004 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4dd89d48-7c16-45d5-9b2c-490fd14d2a74" path="/var/lib/kubelet/pods/4dd89d48-7c16-45d5-9b2c-490fd14d2a74/volumes" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.355390 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.476665 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.565655 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.613968 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.717769 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.798012 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.905850 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ea3ac76-3418-4b1c-8800-d57f554e50e3","Type":"ContainerStarted","Data":"f0754f3ee4636bfe6d2efa2b2933cb347ec873a5b4842b854a3c7378b07c16a6"} Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.907133 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="46554362-2c6d-45b0-b725-e32873db8d22" containerName="manila-share" containerID="cri-o://8073ec086aada98e30c7ecdc5dee9ad2e9cba6368dbf554e3db543698e9587fc" gracePeriod=30 Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.907553 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="46554362-2c6d-45b0-b725-e32873db8d22" containerName="probe" containerID="cri-o://3955a5df80db77c3a24f2787390471f727a6289242ce236c6b0a17d4018fba5a" gracePeriod=30 Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.907600 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="8d192104-a1b7-487b-bfda-72908d31075c" containerName="manila-scheduler" containerID="cri-o://07b9628488431caba14b07b4eeb098576d3f811d9caf39f4ffdfcd3351ac3001" gracePeriod=30 Nov 22 05:46:22 crc kubenswrapper[4818]: I1122 05:46:22.907769 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="8d192104-a1b7-487b-bfda-72908d31075c" containerName="probe" containerID="cri-o://81fe080118a54161335c509815cd015f364ff2ab8dc85613816907f3047783f2" gracePeriod=30 Nov 22 05:46:23 crc kubenswrapper[4818]: E1122 05:46:23.147017 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3759d20_d679_42de_9a4e_f32305ec1ca0.slice/crio-aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3759d20_d679_42de_9a4e_f32305ec1ca0.slice/crio-conmon-aaf1b611dee686f0c296e027ea10a09621aeb3bb14aa4c5b8d991cfb30fe0514.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d192104_a1b7_487b_bfda_72908d31075c.slice/crio-conmon-81fe080118a54161335c509815cd015f364ff2ab8dc85613816907f3047783f2.scope\": RecentStats: unable to find data in memory cache]" Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.830385 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.880424 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gbdz\" (UniqueName: \"kubernetes.io/projected/46554362-2c6d-45b0-b725-e32873db8d22-kube-api-access-9gbdz\") pod \"46554362-2c6d-45b0-b725-e32873db8d22\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.880509 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/46554362-2c6d-45b0-b725-e32873db8d22-var-lib-manila\") pod \"46554362-2c6d-45b0-b725-e32873db8d22\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.880579 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-scripts\") pod \"46554362-2c6d-45b0-b725-e32873db8d22\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.880651 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46554362-2c6d-45b0-b725-e32873db8d22-etc-machine-id\") pod \"46554362-2c6d-45b0-b725-e32873db8d22\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.880713 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-combined-ca-bundle\") pod \"46554362-2c6d-45b0-b725-e32873db8d22\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.880740 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/46554362-2c6d-45b0-b725-e32873db8d22-ceph\") pod \"46554362-2c6d-45b0-b725-e32873db8d22\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.880729 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46554362-2c6d-45b0-b725-e32873db8d22-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "46554362-2c6d-45b0-b725-e32873db8d22" (UID: "46554362-2c6d-45b0-b725-e32873db8d22"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.880819 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-config-data\") pod \"46554362-2c6d-45b0-b725-e32873db8d22\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.880879 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-config-data-custom\") pod \"46554362-2c6d-45b0-b725-e32873db8d22\" (UID: \"46554362-2c6d-45b0-b725-e32873db8d22\") " Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.881403 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46554362-2c6d-45b0-b725-e32873db8d22-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "46554362-2c6d-45b0-b725-e32873db8d22" (UID: "46554362-2c6d-45b0-b725-e32873db8d22"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.881426 4818 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46554362-2c6d-45b0-b725-e32873db8d22-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.886706 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-scripts" (OuterVolumeSpecName: "scripts") pod "46554362-2c6d-45b0-b725-e32873db8d22" (UID: "46554362-2c6d-45b0-b725-e32873db8d22"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.887118 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46554362-2c6d-45b0-b725-e32873db8d22-kube-api-access-9gbdz" (OuterVolumeSpecName: "kube-api-access-9gbdz") pod "46554362-2c6d-45b0-b725-e32873db8d22" (UID: "46554362-2c6d-45b0-b725-e32873db8d22"). InnerVolumeSpecName "kube-api-access-9gbdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.888781 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46554362-2c6d-45b0-b725-e32873db8d22-ceph" (OuterVolumeSpecName: "ceph") pod "46554362-2c6d-45b0-b725-e32873db8d22" (UID: "46554362-2c6d-45b0-b725-e32873db8d22"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.889202 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "46554362-2c6d-45b0-b725-e32873db8d22" (UID: "46554362-2c6d-45b0-b725-e32873db8d22"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.931751 4818 generic.go:334] "Generic (PLEG): container finished" podID="8d192104-a1b7-487b-bfda-72908d31075c" containerID="81fe080118a54161335c509815cd015f364ff2ab8dc85613816907f3047783f2" exitCode=0 Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.931824 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8d192104-a1b7-487b-bfda-72908d31075c","Type":"ContainerDied","Data":"81fe080118a54161335c509815cd015f364ff2ab8dc85613816907f3047783f2"} Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.935469 4818 generic.go:334] "Generic (PLEG): container finished" podID="46554362-2c6d-45b0-b725-e32873db8d22" containerID="3955a5df80db77c3a24f2787390471f727a6289242ce236c6b0a17d4018fba5a" exitCode=0 Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.935514 4818 generic.go:334] "Generic (PLEG): container finished" podID="46554362-2c6d-45b0-b725-e32873db8d22" containerID="8073ec086aada98e30c7ecdc5dee9ad2e9cba6368dbf554e3db543698e9587fc" exitCode=1 Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.935512 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"46554362-2c6d-45b0-b725-e32873db8d22","Type":"ContainerDied","Data":"3955a5df80db77c3a24f2787390471f727a6289242ce236c6b0a17d4018fba5a"} Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.935574 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"46554362-2c6d-45b0-b725-e32873db8d22","Type":"ContainerDied","Data":"8073ec086aada98e30c7ecdc5dee9ad2e9cba6368dbf554e3db543698e9587fc"} Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.935588 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"46554362-2c6d-45b0-b725-e32873db8d22","Type":"ContainerDied","Data":"aa5028bef55a81fc8b30918c1cf15a51fb39367c70864f9390e7c56aaf26b8a1"} Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.935617 4818 scope.go:117] "RemoveContainer" containerID="3955a5df80db77c3a24f2787390471f727a6289242ce236c6b0a17d4018fba5a" Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.935540 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.975965 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46554362-2c6d-45b0-b725-e32873db8d22" (UID: "46554362-2c6d-45b0-b725-e32873db8d22"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.983240 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.983291 4818 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/46554362-2c6d-45b0-b725-e32873db8d22-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.983304 4818 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.983313 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gbdz\" (UniqueName: \"kubernetes.io/projected/46554362-2c6d-45b0-b725-e32873db8d22-kube-api-access-9gbdz\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.983323 4818 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/46554362-2c6d-45b0-b725-e32873db8d22-var-lib-manila\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:23 crc kubenswrapper[4818]: I1122 05:46:23.983330 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.048516 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-config-data" (OuterVolumeSpecName: "config-data") pod "46554362-2c6d-45b0-b725-e32873db8d22" (UID: "46554362-2c6d-45b0-b725-e32873db8d22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.085016 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46554362-2c6d-45b0-b725-e32873db8d22-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.115572 4818 scope.go:117] "RemoveContainer" containerID="8073ec086aada98e30c7ecdc5dee9ad2e9cba6368dbf554e3db543698e9587fc" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.144770 4818 scope.go:117] "RemoveContainer" containerID="3955a5df80db77c3a24f2787390471f727a6289242ce236c6b0a17d4018fba5a" Nov 22 05:46:24 crc kubenswrapper[4818]: E1122 05:46:24.150817 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3955a5df80db77c3a24f2787390471f727a6289242ce236c6b0a17d4018fba5a\": container with ID starting with 3955a5df80db77c3a24f2787390471f727a6289242ce236c6b0a17d4018fba5a not found: ID does not exist" containerID="3955a5df80db77c3a24f2787390471f727a6289242ce236c6b0a17d4018fba5a" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.150869 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3955a5df80db77c3a24f2787390471f727a6289242ce236c6b0a17d4018fba5a"} err="failed to get container status \"3955a5df80db77c3a24f2787390471f727a6289242ce236c6b0a17d4018fba5a\": rpc error: code = NotFound desc = could not find container \"3955a5df80db77c3a24f2787390471f727a6289242ce236c6b0a17d4018fba5a\": container with ID starting with 3955a5df80db77c3a24f2787390471f727a6289242ce236c6b0a17d4018fba5a not found: ID does not exist" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.150901 4818 scope.go:117] "RemoveContainer" containerID="8073ec086aada98e30c7ecdc5dee9ad2e9cba6368dbf554e3db543698e9587fc" Nov 22 05:46:24 crc kubenswrapper[4818]: E1122 05:46:24.151234 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8073ec086aada98e30c7ecdc5dee9ad2e9cba6368dbf554e3db543698e9587fc\": container with ID starting with 8073ec086aada98e30c7ecdc5dee9ad2e9cba6368dbf554e3db543698e9587fc not found: ID does not exist" containerID="8073ec086aada98e30c7ecdc5dee9ad2e9cba6368dbf554e3db543698e9587fc" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.151275 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8073ec086aada98e30c7ecdc5dee9ad2e9cba6368dbf554e3db543698e9587fc"} err="failed to get container status \"8073ec086aada98e30c7ecdc5dee9ad2e9cba6368dbf554e3db543698e9587fc\": rpc error: code = NotFound desc = could not find container \"8073ec086aada98e30c7ecdc5dee9ad2e9cba6368dbf554e3db543698e9587fc\": container with ID starting with 8073ec086aada98e30c7ecdc5dee9ad2e9cba6368dbf554e3db543698e9587fc not found: ID does not exist" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.151292 4818 scope.go:117] "RemoveContainer" containerID="3955a5df80db77c3a24f2787390471f727a6289242ce236c6b0a17d4018fba5a" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.151729 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3955a5df80db77c3a24f2787390471f727a6289242ce236c6b0a17d4018fba5a"} err="failed to get container status \"3955a5df80db77c3a24f2787390471f727a6289242ce236c6b0a17d4018fba5a\": rpc error: code = NotFound desc = could not find container \"3955a5df80db77c3a24f2787390471f727a6289242ce236c6b0a17d4018fba5a\": container with ID starting with 3955a5df80db77c3a24f2787390471f727a6289242ce236c6b0a17d4018fba5a not found: ID does not exist" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.151753 4818 scope.go:117] "RemoveContainer" containerID="8073ec086aada98e30c7ecdc5dee9ad2e9cba6368dbf554e3db543698e9587fc" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.152091 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8073ec086aada98e30c7ecdc5dee9ad2e9cba6368dbf554e3db543698e9587fc"} err="failed to get container status \"8073ec086aada98e30c7ecdc5dee9ad2e9cba6368dbf554e3db543698e9587fc\": rpc error: code = NotFound desc = could not find container \"8073ec086aada98e30c7ecdc5dee9ad2e9cba6368dbf554e3db543698e9587fc\": container with ID starting with 8073ec086aada98e30c7ecdc5dee9ad2e9cba6368dbf554e3db543698e9587fc not found: ID does not exist" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.264366 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.273078 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.288718 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Nov 22 05:46:24 crc kubenswrapper[4818]: E1122 05:46:24.289654 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46554362-2c6d-45b0-b725-e32873db8d22" containerName="probe" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.289688 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="46554362-2c6d-45b0-b725-e32873db8d22" containerName="probe" Nov 22 05:46:24 crc kubenswrapper[4818]: E1122 05:46:24.289771 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46554362-2c6d-45b0-b725-e32873db8d22" containerName="manila-share" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.289784 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="46554362-2c6d-45b0-b725-e32873db8d22" containerName="manila-share" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.290223 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="46554362-2c6d-45b0-b725-e32873db8d22" containerName="manila-share" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.290291 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="46554362-2c6d-45b0-b725-e32873db8d22" containerName="probe" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.292364 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.294375 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.309508 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46554362-2c6d-45b0-b725-e32873db8d22" path="/var/lib/kubelet/pods/46554362-2c6d-45b0-b725-e32873db8d22/volumes" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.311751 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.391557 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5573538-84fb-4f11-9186-0dfd22a2232c-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.391904 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5573538-84fb-4f11-9186-0dfd22a2232c-scripts\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.392059 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5573538-84fb-4f11-9186-0dfd22a2232c-config-data\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.392513 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/a5573538-84fb-4f11-9186-0dfd22a2232c-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.392732 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6bpx\" (UniqueName: \"kubernetes.io/projected/a5573538-84fb-4f11-9186-0dfd22a2232c-kube-api-access-z6bpx\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.392880 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5573538-84fb-4f11-9186-0dfd22a2232c-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.393078 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5573538-84fb-4f11-9186-0dfd22a2232c-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.393159 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a5573538-84fb-4f11-9186-0dfd22a2232c-ceph\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.494995 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/a5573538-84fb-4f11-9186-0dfd22a2232c-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.495153 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/a5573538-84fb-4f11-9186-0dfd22a2232c-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.496824 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6bpx\" (UniqueName: \"kubernetes.io/projected/a5573538-84fb-4f11-9186-0dfd22a2232c-kube-api-access-z6bpx\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.497182 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5573538-84fb-4f11-9186-0dfd22a2232c-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.497626 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5573538-84fb-4f11-9186-0dfd22a2232c-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.498138 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5573538-84fb-4f11-9186-0dfd22a2232c-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.499887 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a5573538-84fb-4f11-9186-0dfd22a2232c-ceph\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.500202 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5573538-84fb-4f11-9186-0dfd22a2232c-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.500456 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5573538-84fb-4f11-9186-0dfd22a2232c-scripts\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.501085 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5573538-84fb-4f11-9186-0dfd22a2232c-config-data\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.504538 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a5573538-84fb-4f11-9186-0dfd22a2232c-ceph\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.505672 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5573538-84fb-4f11-9186-0dfd22a2232c-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.506055 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5573538-84fb-4f11-9186-0dfd22a2232c-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.509738 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5573538-84fb-4f11-9186-0dfd22a2232c-config-data\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.513828 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5573538-84fb-4f11-9186-0dfd22a2232c-scripts\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.518568 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6bpx\" (UniqueName: \"kubernetes.io/projected/a5573538-84fb-4f11-9186-0dfd22a2232c-kube-api-access-z6bpx\") pod \"manila-share-share1-0\" (UID: \"a5573538-84fb-4f11-9186-0dfd22a2232c\") " pod="openstack/manila-share-share1-0" Nov 22 05:46:24 crc kubenswrapper[4818]: I1122 05:46:24.618187 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 22 05:46:25 crc kubenswrapper[4818]: I1122 05:46:24.948815 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ea3ac76-3418-4b1c-8800-d57f554e50e3","Type":"ContainerStarted","Data":"67a899034d98575a63696c14cd3dd1462fc62dba474761961d6638d15bf4ae73"} Nov 22 05:46:25 crc kubenswrapper[4818]: I1122 05:46:24.949079 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ea3ac76-3418-4b1c-8800-d57f554e50e3","Type":"ContainerStarted","Data":"20c34089231828b63ea4546e467d13b8ed90629eb4b7d547c51cffebe377437a"} Nov 22 05:46:25 crc kubenswrapper[4818]: I1122 05:46:25.854610 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 22 05:46:25 crc kubenswrapper[4818]: I1122 05:46:25.962357 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"a5573538-84fb-4f11-9186-0dfd22a2232c","Type":"ContainerStarted","Data":"8e2a52fa3dc13fe0fb42d919cfe6a9c11c667477ec37bd5e6a2c2c61a7b9c79a"} Nov 22 05:46:25 crc kubenswrapper[4818]: I1122 05:46:25.963987 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ea3ac76-3418-4b1c-8800-d57f554e50e3","Type":"ContainerStarted","Data":"d2e34379c2a5504c76b14e022b69f0ca3412af7a7ba5e498fb8137410cc93eeb"} Nov 22 05:46:26 crc kubenswrapper[4818]: I1122 05:46:26.975349 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ea3ac76-3418-4b1c-8800-d57f554e50e3","Type":"ContainerStarted","Data":"2ca1caf8a0be210c1873cdae8290d56f18312c7d04963aafd4c522e85fbd7cba"} Nov 22 05:46:26 crc kubenswrapper[4818]: I1122 05:46:26.983651 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 05:46:26 crc kubenswrapper[4818]: I1122 05:46:26.983830 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"a5573538-84fb-4f11-9186-0dfd22a2232c","Type":"ContainerStarted","Data":"c09a5b03c71a3fa3d7bc5a2d8773b974aeefc2cb58dee03dc4896e69eff49ba3"} Nov 22 05:46:26 crc kubenswrapper[4818]: I1122 05:46:26.983950 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"a5573538-84fb-4f11-9186-0dfd22a2232c","Type":"ContainerStarted","Data":"faecac2d81efac40a25ddf4de83590836751a88e09d7fc0ccc9d073079d668e2"} Nov 22 05:46:26 crc kubenswrapper[4818]: I1122 05:46:26.994951 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.141804179 podStartE2EDuration="5.994937473s" podCreationTimestamp="2025-11-22 05:46:21 +0000 UTC" firstStartedPulling="2025-11-22 05:46:22.617499642 +0000 UTC m=+3535.191916169" lastFinishedPulling="2025-11-22 05:46:26.470632916 +0000 UTC m=+3539.045049463" observedRunningTime="2025-11-22 05:46:26.991962114 +0000 UTC m=+3539.566378641" watchObservedRunningTime="2025-11-22 05:46:26.994937473 +0000 UTC m=+3539.569354000" Nov 22 05:46:27 crc kubenswrapper[4818]: I1122 05:46:27.042588 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.042570233 podStartE2EDuration="3.042570233s" podCreationTimestamp="2025-11-22 05:46:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:46:27.03722468 +0000 UTC m=+3539.611641237" watchObservedRunningTime="2025-11-22 05:46:27.042570233 +0000 UTC m=+3539.616986760" Nov 22 05:46:27 crc kubenswrapper[4818]: I1122 05:46:27.964809 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 22 05:46:27 crc kubenswrapper[4818]: I1122 05:46:27.986991 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d192104-a1b7-487b-bfda-72908d31075c-etc-machine-id\") pod \"8d192104-a1b7-487b-bfda-72908d31075c\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " Nov 22 05:46:27 crc kubenswrapper[4818]: I1122 05:46:27.987054 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-config-data-custom\") pod \"8d192104-a1b7-487b-bfda-72908d31075c\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " Nov 22 05:46:27 crc kubenswrapper[4818]: I1122 05:46:27.987098 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-config-data\") pod \"8d192104-a1b7-487b-bfda-72908d31075c\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " Nov 22 05:46:27 crc kubenswrapper[4818]: I1122 05:46:27.987114 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8d192104-a1b7-487b-bfda-72908d31075c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8d192104-a1b7-487b-bfda-72908d31075c" (UID: "8d192104-a1b7-487b-bfda-72908d31075c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 05:46:27 crc kubenswrapper[4818]: I1122 05:46:27.987176 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-combined-ca-bundle\") pod \"8d192104-a1b7-487b-bfda-72908d31075c\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " Nov 22 05:46:27 crc kubenswrapper[4818]: I1122 05:46:27.987302 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-scripts\") pod \"8d192104-a1b7-487b-bfda-72908d31075c\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " Nov 22 05:46:27 crc kubenswrapper[4818]: I1122 05:46:27.987454 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8cd9\" (UniqueName: \"kubernetes.io/projected/8d192104-a1b7-487b-bfda-72908d31075c-kube-api-access-j8cd9\") pod \"8d192104-a1b7-487b-bfda-72908d31075c\" (UID: \"8d192104-a1b7-487b-bfda-72908d31075c\") " Nov 22 05:46:27 crc kubenswrapper[4818]: I1122 05:46:27.987965 4818 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d192104-a1b7-487b-bfda-72908d31075c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:27 crc kubenswrapper[4818]: I1122 05:46:27.989805 4818 generic.go:334] "Generic (PLEG): container finished" podID="8d192104-a1b7-487b-bfda-72908d31075c" containerID="07b9628488431caba14b07b4eeb098576d3f811d9caf39f4ffdfcd3351ac3001" exitCode=0 Nov 22 05:46:27 crc kubenswrapper[4818]: I1122 05:46:27.992755 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8d192104-a1b7-487b-bfda-72908d31075c" (UID: "8d192104-a1b7-487b-bfda-72908d31075c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:27 crc kubenswrapper[4818]: I1122 05:46:27.992867 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 22 05:46:27 crc kubenswrapper[4818]: I1122 05:46:27.993605 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8d192104-a1b7-487b-bfda-72908d31075c","Type":"ContainerDied","Data":"07b9628488431caba14b07b4eeb098576d3f811d9caf39f4ffdfcd3351ac3001"} Nov 22 05:46:27 crc kubenswrapper[4818]: I1122 05:46:27.993643 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8d192104-a1b7-487b-bfda-72908d31075c","Type":"ContainerDied","Data":"623fe0080210daf12ec0904c4c93aec009ab8f41de24e7b5c56bab49321c0ed3"} Nov 22 05:46:27 crc kubenswrapper[4818]: I1122 05:46:27.993665 4818 scope.go:117] "RemoveContainer" containerID="81fe080118a54161335c509815cd015f364ff2ab8dc85613816907f3047783f2" Nov 22 05:46:27 crc kubenswrapper[4818]: I1122 05:46:27.993745 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-scripts" (OuterVolumeSpecName: "scripts") pod "8d192104-a1b7-487b-bfda-72908d31075c" (UID: "8d192104-a1b7-487b-bfda-72908d31075c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.005078 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d192104-a1b7-487b-bfda-72908d31075c-kube-api-access-j8cd9" (OuterVolumeSpecName: "kube-api-access-j8cd9") pod "8d192104-a1b7-487b-bfda-72908d31075c" (UID: "8d192104-a1b7-487b-bfda-72908d31075c"). InnerVolumeSpecName "kube-api-access-j8cd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.073421 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d192104-a1b7-487b-bfda-72908d31075c" (UID: "8d192104-a1b7-487b-bfda-72908d31075c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.092841 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8cd9\" (UniqueName: \"kubernetes.io/projected/8d192104-a1b7-487b-bfda-72908d31075c-kube-api-access-j8cd9\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.092880 4818 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.092903 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.092917 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.113076 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-config-data" (OuterVolumeSpecName: "config-data") pod "8d192104-a1b7-487b-bfda-72908d31075c" (UID: "8d192104-a1b7-487b-bfda-72908d31075c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.143671 4818 scope.go:117] "RemoveContainer" containerID="07b9628488431caba14b07b4eeb098576d3f811d9caf39f4ffdfcd3351ac3001" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.163277 4818 scope.go:117] "RemoveContainer" containerID="81fe080118a54161335c509815cd015f364ff2ab8dc85613816907f3047783f2" Nov 22 05:46:28 crc kubenswrapper[4818]: E1122 05:46:28.163659 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81fe080118a54161335c509815cd015f364ff2ab8dc85613816907f3047783f2\": container with ID starting with 81fe080118a54161335c509815cd015f364ff2ab8dc85613816907f3047783f2 not found: ID does not exist" containerID="81fe080118a54161335c509815cd015f364ff2ab8dc85613816907f3047783f2" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.163699 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81fe080118a54161335c509815cd015f364ff2ab8dc85613816907f3047783f2"} err="failed to get container status \"81fe080118a54161335c509815cd015f364ff2ab8dc85613816907f3047783f2\": rpc error: code = NotFound desc = could not find container \"81fe080118a54161335c509815cd015f364ff2ab8dc85613816907f3047783f2\": container with ID starting with 81fe080118a54161335c509815cd015f364ff2ab8dc85613816907f3047783f2 not found: ID does not exist" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.163729 4818 scope.go:117] "RemoveContainer" containerID="07b9628488431caba14b07b4eeb098576d3f811d9caf39f4ffdfcd3351ac3001" Nov 22 05:46:28 crc kubenswrapper[4818]: E1122 05:46:28.163968 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07b9628488431caba14b07b4eeb098576d3f811d9caf39f4ffdfcd3351ac3001\": container with ID starting with 07b9628488431caba14b07b4eeb098576d3f811d9caf39f4ffdfcd3351ac3001 not found: ID does not exist" containerID="07b9628488431caba14b07b4eeb098576d3f811d9caf39f4ffdfcd3351ac3001" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.163999 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07b9628488431caba14b07b4eeb098576d3f811d9caf39f4ffdfcd3351ac3001"} err="failed to get container status \"07b9628488431caba14b07b4eeb098576d3f811d9caf39f4ffdfcd3351ac3001\": rpc error: code = NotFound desc = could not find container \"07b9628488431caba14b07b4eeb098576d3f811d9caf39f4ffdfcd3351ac3001\": container with ID starting with 07b9628488431caba14b07b4eeb098576d3f811d9caf39f4ffdfcd3351ac3001 not found: ID does not exist" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.195033 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d192104-a1b7-487b-bfda-72908d31075c-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.353132 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.368310 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.407742 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Nov 22 05:46:28 crc kubenswrapper[4818]: E1122 05:46:28.409414 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d192104-a1b7-487b-bfda-72908d31075c" containerName="manila-scheduler" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.409464 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d192104-a1b7-487b-bfda-72908d31075c" containerName="manila-scheduler" Nov 22 05:46:28 crc kubenswrapper[4818]: E1122 05:46:28.409537 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d192104-a1b7-487b-bfda-72908d31075c" containerName="probe" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.409547 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d192104-a1b7-487b-bfda-72908d31075c" containerName="probe" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.410278 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d192104-a1b7-487b-bfda-72908d31075c" containerName="manila-scheduler" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.410321 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d192104-a1b7-487b-bfda-72908d31075c" containerName="probe" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.413194 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.415902 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.421296 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 22 05:46:28 crc kubenswrapper[4818]: E1122 05:46:28.434607 4818 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/b478e1daf18ec2464e7365e133fd0fae97df64bd4eaac43172b662ad098682fc/diff" to get inode usage: stat /var/lib/containers/storage/overlay/b478e1daf18ec2464e7365e133fd0fae97df64bd4eaac43172b662ad098682fc/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openshift-marketplace_certified-operators-ksfkm_e3759d20-d679-42de-9a4e-f32305ec1ca0/extract-utilities/0.log" to get inode usage: stat /var/log/pods/openshift-marketplace_certified-operators-ksfkm_e3759d20-d679-42de-9a4e-f32305ec1ca0/extract-utilities/0.log: no such file or directory Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.506120 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.506286 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.506343 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a-config-data\") pod \"manila-scheduler-0\" (UID: \"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.506383 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.506721 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs4jb\" (UniqueName: \"kubernetes.io/projected/fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a-kube-api-access-zs4jb\") pod \"manila-scheduler-0\" (UID: \"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.506821 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a-scripts\") pod \"manila-scheduler-0\" (UID: \"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.539729 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.608490 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.608623 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs4jb\" (UniqueName: \"kubernetes.io/projected/fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a-kube-api-access-zs4jb\") pod \"manila-scheduler-0\" (UID: \"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.608650 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a-scripts\") pod \"manila-scheduler-0\" (UID: \"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.608688 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.608761 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.608793 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a-config-data\") pod \"manila-scheduler-0\" (UID: \"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.612525 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.614435 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a-config-data\") pod \"manila-scheduler-0\" (UID: \"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.617896 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.618456 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a-scripts\") pod \"manila-scheduler-0\" (UID: \"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.619160 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.633243 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs4jb\" (UniqueName: \"kubernetes.io/projected/fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a-kube-api-access-zs4jb\") pod \"manila-scheduler-0\" (UID: \"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a\") " pod="openstack/manila-scheduler-0" Nov 22 05:46:28 crc kubenswrapper[4818]: I1122 05:46:28.730813 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 22 05:46:29 crc kubenswrapper[4818]: I1122 05:46:29.025695 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 22 05:46:30 crc kubenswrapper[4818]: I1122 05:46:30.014473 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a","Type":"ContainerStarted","Data":"2be9f0424be7088e90dd90de6a9a1d52e3e45e7f5f83adf380e7506673375c8a"} Nov 22 05:46:30 crc kubenswrapper[4818]: I1122 05:46:30.015021 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a","Type":"ContainerStarted","Data":"0ddb753fa36bc534065d27189d1d837858419c1eebdc724ef68821d5cb874f07"} Nov 22 05:46:30 crc kubenswrapper[4818]: I1122 05:46:30.312644 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d192104-a1b7-487b-bfda-72908d31075c" path="/var/lib/kubelet/pods/8d192104-a1b7-487b-bfda-72908d31075c/volumes" Nov 22 05:46:31 crc kubenswrapper[4818]: I1122 05:46:31.031381 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a","Type":"ContainerStarted","Data":"c51f7f39a14ff65cb882054c8f9255a2b3c7a792dccc5f66cd0c0efae301bcc2"} Nov 22 05:46:31 crc kubenswrapper[4818]: I1122 05:46:31.066853 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.066833449 podStartE2EDuration="3.066833449s" podCreationTimestamp="2025-11-22 05:46:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 05:46:31.055430666 +0000 UTC m=+3543.629847213" watchObservedRunningTime="2025-11-22 05:46:31.066833449 +0000 UTC m=+3543.641249986" Nov 22 05:46:31 crc kubenswrapper[4818]: I1122 05:46:31.102338 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7764859bb4-tnzkk" podUID="60b9da36-1281-4cc5-a9ac-a2df2e4d9638" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.242:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.242:8443: connect: connection refused" Nov 22 05:46:34 crc kubenswrapper[4818]: I1122 05:46:34.619632 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Nov 22 05:46:38 crc kubenswrapper[4818]: I1122 05:46:38.730939 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Nov 22 05:46:41 crc kubenswrapper[4818]: I1122 05:46:41.102640 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7764859bb4-tnzkk" podUID="60b9da36-1281-4cc5-a9ac-a2df2e4d9638" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.242:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.242:8443: connect: connection refused" Nov 22 05:46:41 crc kubenswrapper[4818]: I1122 05:46:41.103148 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:46:46 crc kubenswrapper[4818]: E1122 05:46:46.122461 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60b9da36_1281_4cc5_a9ac_a2df2e4d9638.slice/crio-conmon-180e7581aa51496380128694e8a8eda047b65550ed4a7f9e238af15906922636.scope\": RecentStats: unable to find data in memory cache]" Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.132544 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.213184 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7764859bb4-tnzkk" event={"ID":"60b9da36-1281-4cc5-a9ac-a2df2e4d9638","Type":"ContainerDied","Data":"180e7581aa51496380128694e8a8eda047b65550ed4a7f9e238af15906922636"} Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.213125 4818 generic.go:334] "Generic (PLEG): container finished" podID="60b9da36-1281-4cc5-a9ac-a2df2e4d9638" containerID="180e7581aa51496380128694e8a8eda047b65550ed4a7f9e238af15906922636" exitCode=137 Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.213747 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7764859bb4-tnzkk" event={"ID":"60b9da36-1281-4cc5-a9ac-a2df2e4d9638","Type":"ContainerDied","Data":"7c11e3c6df9c02580c092fb06c3874060f2efafbabbc374db2914c8fd2166d13"} Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.213843 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c11e3c6df9c02580c092fb06c3874060f2efafbabbc374db2914c8fd2166d13" Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.266498 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.270427 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-logs\") pod \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.270481 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-config-data\") pod \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.270518 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-horizon-tls-certs\") pod \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.270541 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-combined-ca-bundle\") pod \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.270564 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-scripts\") pod \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.270619 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tsmx\" (UniqueName: \"kubernetes.io/projected/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-kube-api-access-7tsmx\") pod \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.270662 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-horizon-secret-key\") pod \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\" (UID: \"60b9da36-1281-4cc5-a9ac-a2df2e4d9638\") " Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.271050 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-logs" (OuterVolumeSpecName: "logs") pod "60b9da36-1281-4cc5-a9ac-a2df2e4d9638" (UID: "60b9da36-1281-4cc5-a9ac-a2df2e4d9638"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.277143 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-kube-api-access-7tsmx" (OuterVolumeSpecName: "kube-api-access-7tsmx") pod "60b9da36-1281-4cc5-a9ac-a2df2e4d9638" (UID: "60b9da36-1281-4cc5-a9ac-a2df2e4d9638"). InnerVolumeSpecName "kube-api-access-7tsmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.277936 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "60b9da36-1281-4cc5-a9ac-a2df2e4d9638" (UID: "60b9da36-1281-4cc5-a9ac-a2df2e4d9638"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.313070 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-config-data" (OuterVolumeSpecName: "config-data") pod "60b9da36-1281-4cc5-a9ac-a2df2e4d9638" (UID: "60b9da36-1281-4cc5-a9ac-a2df2e4d9638"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.328210 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-scripts" (OuterVolumeSpecName: "scripts") pod "60b9da36-1281-4cc5-a9ac-a2df2e4d9638" (UID: "60b9da36-1281-4cc5-a9ac-a2df2e4d9638"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.334574 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60b9da36-1281-4cc5-a9ac-a2df2e4d9638" (UID: "60b9da36-1281-4cc5-a9ac-a2df2e4d9638"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.338805 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "60b9da36-1281-4cc5-a9ac-a2df2e4d9638" (UID: "60b9da36-1281-4cc5-a9ac-a2df2e4d9638"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.375755 4818 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.375800 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.375857 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.376111 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tsmx\" (UniqueName: \"kubernetes.io/projected/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-kube-api-access-7tsmx\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.376130 4818 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.376138 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-logs\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:46 crc kubenswrapper[4818]: I1122 05:46:46.376146 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60b9da36-1281-4cc5-a9ac-a2df2e4d9638-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 05:46:47 crc kubenswrapper[4818]: I1122 05:46:47.222937 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7764859bb4-tnzkk" Nov 22 05:46:47 crc kubenswrapper[4818]: I1122 05:46:47.265737 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7764859bb4-tnzkk"] Nov 22 05:46:47 crc kubenswrapper[4818]: I1122 05:46:47.280209 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7764859bb4-tnzkk"] Nov 22 05:46:48 crc kubenswrapper[4818]: I1122 05:46:48.304370 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60b9da36-1281-4cc5-a9ac-a2df2e4d9638" path="/var/lib/kubelet/pods/60b9da36-1281-4cc5-a9ac-a2df2e4d9638/volumes" Nov 22 05:46:50 crc kubenswrapper[4818]: I1122 05:46:50.183750 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Nov 22 05:46:51 crc kubenswrapper[4818]: I1122 05:46:51.265200 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:46:51 crc kubenswrapper[4818]: I1122 05:46:51.265361 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:46:51 crc kubenswrapper[4818]: I1122 05:46:51.265426 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 05:46:51 crc kubenswrapper[4818]: I1122 05:46:51.266519 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"30eba2944bc07de93901d264862a9a44b1afc7503d2ddd03ddaec3e1e784bc41"} pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 05:46:51 crc kubenswrapper[4818]: I1122 05:46:51.266628 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" containerID="cri-o://30eba2944bc07de93901d264862a9a44b1afc7503d2ddd03ddaec3e1e784bc41" gracePeriod=600 Nov 22 05:46:52 crc kubenswrapper[4818]: I1122 05:46:52.272669 4818 generic.go:334] "Generic (PLEG): container finished" podID="0971c9ec-29ce-44eb-9875-ad4664143305" containerID="30eba2944bc07de93901d264862a9a44b1afc7503d2ddd03ddaec3e1e784bc41" exitCode=0 Nov 22 05:46:52 crc kubenswrapper[4818]: I1122 05:46:52.272745 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerDied","Data":"30eba2944bc07de93901d264862a9a44b1afc7503d2ddd03ddaec3e1e784bc41"} Nov 22 05:46:52 crc kubenswrapper[4818]: I1122 05:46:52.273267 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849"} Nov 22 05:46:52 crc kubenswrapper[4818]: I1122 05:46:52.273292 4818 scope.go:117] "RemoveContainer" containerID="297c99cdc2ebeac2c8dee35807c20d7dd41aab410eb5d37a33e8dd7a0114598a" Nov 22 05:46:52 crc kubenswrapper[4818]: I1122 05:46:52.378723 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.137851 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 22 05:47:43 crc kubenswrapper[4818]: E1122 05:47:43.139250 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60b9da36-1281-4cc5-a9ac-a2df2e4d9638" containerName="horizon" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.139299 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="60b9da36-1281-4cc5-a9ac-a2df2e4d9638" containerName="horizon" Nov 22 05:47:43 crc kubenswrapper[4818]: E1122 05:47:43.139341 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60b9da36-1281-4cc5-a9ac-a2df2e4d9638" containerName="horizon-log" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.139353 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="60b9da36-1281-4cc5-a9ac-a2df2e4d9638" containerName="horizon-log" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.139650 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="60b9da36-1281-4cc5-a9ac-a2df2e4d9638" containerName="horizon-log" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.139686 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="60b9da36-1281-4cc5-a9ac-a2df2e4d9638" containerName="horizon" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.140783 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.143687 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-9vz8j" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.148970 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.149026 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.149119 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.149681 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.272521 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/df264e71-c1de-4e1f-9a56-24f5bd284a80-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.272596 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.272642 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/df264e71-c1de-4e1f-9a56-24f5bd284a80-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.273005 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df264e71-c1de-4e1f-9a56-24f5bd284a80-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.273453 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/df264e71-c1de-4e1f-9a56-24f5bd284a80-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.273557 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/df264e71-c1de-4e1f-9a56-24f5bd284a80-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.273646 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw2nn\" (UniqueName: \"kubernetes.io/projected/df264e71-c1de-4e1f-9a56-24f5bd284a80-kube-api-access-cw2nn\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.273704 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df264e71-c1de-4e1f-9a56-24f5bd284a80-config-data\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.273766 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/df264e71-c1de-4e1f-9a56-24f5bd284a80-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.376062 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df264e71-c1de-4e1f-9a56-24f5bd284a80-config-data\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.376514 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/df264e71-c1de-4e1f-9a56-24f5bd284a80-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.376579 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/df264e71-c1de-4e1f-9a56-24f5bd284a80-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.376623 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.376666 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/df264e71-c1de-4e1f-9a56-24f5bd284a80-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.376693 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df264e71-c1de-4e1f-9a56-24f5bd284a80-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.376798 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/df264e71-c1de-4e1f-9a56-24f5bd284a80-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.376848 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/df264e71-c1de-4e1f-9a56-24f5bd284a80-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.376877 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw2nn\" (UniqueName: \"kubernetes.io/projected/df264e71-c1de-4e1f-9a56-24f5bd284a80-kube-api-access-cw2nn\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.378191 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.380040 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/df264e71-c1de-4e1f-9a56-24f5bd284a80-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.380512 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df264e71-c1de-4e1f-9a56-24f5bd284a80-config-data\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.380522 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/df264e71-c1de-4e1f-9a56-24f5bd284a80-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.380787 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/df264e71-c1de-4e1f-9a56-24f5bd284a80-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.387049 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/df264e71-c1de-4e1f-9a56-24f5bd284a80-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.388010 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df264e71-c1de-4e1f-9a56-24f5bd284a80-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.388119 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/df264e71-c1de-4e1f-9a56-24f5bd284a80-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.404133 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw2nn\" (UniqueName: \"kubernetes.io/projected/df264e71-c1de-4e1f-9a56-24f5bd284a80-kube-api-access-cw2nn\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.412337 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.470781 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 22 05:47:43 crc kubenswrapper[4818]: I1122 05:47:43.974287 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 22 05:47:44 crc kubenswrapper[4818]: I1122 05:47:44.892386 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"df264e71-c1de-4e1f-9a56-24f5bd284a80","Type":"ContainerStarted","Data":"ef0e41d6ee637b3a795992b3238c7cdec4790c79ff7269354ce57917c414deab"} Nov 22 05:48:11 crc kubenswrapper[4818]: E1122 05:48:11.267501 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 22 05:48:11 crc kubenswrapper[4818]: E1122 05:48:11.268775 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cw2nn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(df264e71-c1de-4e1f-9a56-24f5bd284a80): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 05:48:11 crc kubenswrapper[4818]: E1122 05:48:11.270124 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="df264e71-c1de-4e1f-9a56-24f5bd284a80" Nov 22 05:48:12 crc kubenswrapper[4818]: E1122 05:48:12.146934 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="df264e71-c1de-4e1f-9a56-24f5bd284a80" Nov 22 05:48:25 crc kubenswrapper[4818]: I1122 05:48:25.295807 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"df264e71-c1de-4e1f-9a56-24f5bd284a80","Type":"ContainerStarted","Data":"938560beedaa5d2f146e233c4f8cdd631f93c381aba0bfe9cf75146edf77c60f"} Nov 22 05:48:25 crc kubenswrapper[4818]: I1122 05:48:25.331294 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.387049679 podStartE2EDuration="43.331269997s" podCreationTimestamp="2025-11-22 05:47:42 +0000 UTC" firstStartedPulling="2025-11-22 05:47:43.982911939 +0000 UTC m=+3616.557328466" lastFinishedPulling="2025-11-22 05:48:23.927132247 +0000 UTC m=+3656.501548784" observedRunningTime="2025-11-22 05:48:25.326970853 +0000 UTC m=+3657.901387420" watchObservedRunningTime="2025-11-22 05:48:25.331269997 +0000 UTC m=+3657.905686534" Nov 22 05:48:51 crc kubenswrapper[4818]: I1122 05:48:51.264539 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:48:51 crc kubenswrapper[4818]: I1122 05:48:51.265486 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:49:21 crc kubenswrapper[4818]: I1122 05:49:21.265459 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:49:21 crc kubenswrapper[4818]: I1122 05:49:21.266152 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:49:51 crc kubenswrapper[4818]: I1122 05:49:51.264945 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:49:51 crc kubenswrapper[4818]: I1122 05:49:51.266115 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:49:51 crc kubenswrapper[4818]: I1122 05:49:51.266211 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 05:49:51 crc kubenswrapper[4818]: I1122 05:49:51.267689 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849"} pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 05:49:51 crc kubenswrapper[4818]: I1122 05:49:51.267761 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" containerID="cri-o://5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" gracePeriod=600 Nov 22 05:49:51 crc kubenswrapper[4818]: E1122 05:49:51.447534 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:49:51 crc kubenswrapper[4818]: I1122 05:49:51.474723 4818 generic.go:334] "Generic (PLEG): container finished" podID="0971c9ec-29ce-44eb-9875-ad4664143305" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" exitCode=0 Nov 22 05:49:51 crc kubenswrapper[4818]: I1122 05:49:51.474779 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerDied","Data":"5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849"} Nov 22 05:49:51 crc kubenswrapper[4818]: I1122 05:49:51.474819 4818 scope.go:117] "RemoveContainer" containerID="30eba2944bc07de93901d264862a9a44b1afc7503d2ddd03ddaec3e1e784bc41" Nov 22 05:49:51 crc kubenswrapper[4818]: I1122 05:49:51.475987 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:49:51 crc kubenswrapper[4818]: E1122 05:49:51.476893 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:50:02 crc kubenswrapper[4818]: I1122 05:50:02.291093 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:50:02 crc kubenswrapper[4818]: E1122 05:50:02.293501 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:50:16 crc kubenswrapper[4818]: I1122 05:50:16.291972 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:50:16 crc kubenswrapper[4818]: E1122 05:50:16.293385 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:50:28 crc kubenswrapper[4818]: I1122 05:50:28.297599 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:50:28 crc kubenswrapper[4818]: E1122 05:50:28.298784 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:50:41 crc kubenswrapper[4818]: I1122 05:50:41.291908 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:50:41 crc kubenswrapper[4818]: E1122 05:50:41.293047 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:50:54 crc kubenswrapper[4818]: I1122 05:50:54.308280 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:50:54 crc kubenswrapper[4818]: E1122 05:50:54.309183 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:51:08 crc kubenswrapper[4818]: I1122 05:51:08.300737 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:51:08 crc kubenswrapper[4818]: E1122 05:51:08.301547 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:51:22 crc kubenswrapper[4818]: I1122 05:51:22.291914 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:51:22 crc kubenswrapper[4818]: E1122 05:51:22.293157 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:51:36 crc kubenswrapper[4818]: I1122 05:51:36.292273 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:51:36 crc kubenswrapper[4818]: E1122 05:51:36.293466 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:51:43 crc kubenswrapper[4818]: I1122 05:51:43.725164 4818 scope.go:117] "RemoveContainer" containerID="180e7581aa51496380128694e8a8eda047b65550ed4a7f9e238af15906922636" Nov 22 05:51:50 crc kubenswrapper[4818]: I1122 05:51:50.292237 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:51:50 crc kubenswrapper[4818]: E1122 05:51:50.293386 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:52:01 crc kubenswrapper[4818]: I1122 05:52:01.291704 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:52:01 crc kubenswrapper[4818]: E1122 05:52:01.292555 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:52:16 crc kubenswrapper[4818]: I1122 05:52:16.291162 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:52:16 crc kubenswrapper[4818]: E1122 05:52:16.291988 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:52:31 crc kubenswrapper[4818]: I1122 05:52:31.291793 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:52:31 crc kubenswrapper[4818]: E1122 05:52:31.292538 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:52:43 crc kubenswrapper[4818]: I1122 05:52:43.771462 4818 scope.go:117] "RemoveContainer" containerID="64ff53e988010bba9bf54bb360558aabbc0e90386763b14c6ab5fe8f007de1e8" Nov 22 05:52:44 crc kubenswrapper[4818]: I1122 05:52:44.291544 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:52:44 crc kubenswrapper[4818]: E1122 05:52:44.292137 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:52:57 crc kubenswrapper[4818]: I1122 05:52:57.291093 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:52:57 crc kubenswrapper[4818]: E1122 05:52:57.292025 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:53:12 crc kubenswrapper[4818]: I1122 05:53:12.292151 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:53:12 crc kubenswrapper[4818]: E1122 05:53:12.293446 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:53:23 crc kubenswrapper[4818]: I1122 05:53:23.291024 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:53:23 crc kubenswrapper[4818]: E1122 05:53:23.292016 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:53:37 crc kubenswrapper[4818]: I1122 05:53:37.292443 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:53:37 crc kubenswrapper[4818]: E1122 05:53:37.293216 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:53:38 crc kubenswrapper[4818]: I1122 05:53:38.581498 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wp6hh"] Nov 22 05:53:38 crc kubenswrapper[4818]: I1122 05:53:38.583957 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wp6hh" Nov 22 05:53:38 crc kubenswrapper[4818]: I1122 05:53:38.599599 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wp6hh"] Nov 22 05:53:38 crc kubenswrapper[4818]: I1122 05:53:38.667285 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebcedfbe-da71-4e94-af5d-2fc3163fede9-catalog-content\") pod \"redhat-operators-wp6hh\" (UID: \"ebcedfbe-da71-4e94-af5d-2fc3163fede9\") " pod="openshift-marketplace/redhat-operators-wp6hh" Nov 22 05:53:38 crc kubenswrapper[4818]: I1122 05:53:38.667399 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxfn9\" (UniqueName: \"kubernetes.io/projected/ebcedfbe-da71-4e94-af5d-2fc3163fede9-kube-api-access-zxfn9\") pod \"redhat-operators-wp6hh\" (UID: \"ebcedfbe-da71-4e94-af5d-2fc3163fede9\") " pod="openshift-marketplace/redhat-operators-wp6hh" Nov 22 05:53:38 crc kubenswrapper[4818]: I1122 05:53:38.667422 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebcedfbe-da71-4e94-af5d-2fc3163fede9-utilities\") pod \"redhat-operators-wp6hh\" (UID: \"ebcedfbe-da71-4e94-af5d-2fc3163fede9\") " pod="openshift-marketplace/redhat-operators-wp6hh" Nov 22 05:53:38 crc kubenswrapper[4818]: I1122 05:53:38.768853 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxfn9\" (UniqueName: \"kubernetes.io/projected/ebcedfbe-da71-4e94-af5d-2fc3163fede9-kube-api-access-zxfn9\") pod \"redhat-operators-wp6hh\" (UID: \"ebcedfbe-da71-4e94-af5d-2fc3163fede9\") " pod="openshift-marketplace/redhat-operators-wp6hh" Nov 22 05:53:38 crc kubenswrapper[4818]: I1122 05:53:38.768897 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebcedfbe-da71-4e94-af5d-2fc3163fede9-utilities\") pod \"redhat-operators-wp6hh\" (UID: \"ebcedfbe-da71-4e94-af5d-2fc3163fede9\") " pod="openshift-marketplace/redhat-operators-wp6hh" Nov 22 05:53:38 crc kubenswrapper[4818]: I1122 05:53:38.768988 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebcedfbe-da71-4e94-af5d-2fc3163fede9-catalog-content\") pod \"redhat-operators-wp6hh\" (UID: \"ebcedfbe-da71-4e94-af5d-2fc3163fede9\") " pod="openshift-marketplace/redhat-operators-wp6hh" Nov 22 05:53:38 crc kubenswrapper[4818]: I1122 05:53:38.769482 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebcedfbe-da71-4e94-af5d-2fc3163fede9-catalog-content\") pod \"redhat-operators-wp6hh\" (UID: \"ebcedfbe-da71-4e94-af5d-2fc3163fede9\") " pod="openshift-marketplace/redhat-operators-wp6hh" Nov 22 05:53:38 crc kubenswrapper[4818]: I1122 05:53:38.769950 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebcedfbe-da71-4e94-af5d-2fc3163fede9-utilities\") pod \"redhat-operators-wp6hh\" (UID: \"ebcedfbe-da71-4e94-af5d-2fc3163fede9\") " pod="openshift-marketplace/redhat-operators-wp6hh" Nov 22 05:53:38 crc kubenswrapper[4818]: I1122 05:53:38.802486 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxfn9\" (UniqueName: \"kubernetes.io/projected/ebcedfbe-da71-4e94-af5d-2fc3163fede9-kube-api-access-zxfn9\") pod \"redhat-operators-wp6hh\" (UID: \"ebcedfbe-da71-4e94-af5d-2fc3163fede9\") " pod="openshift-marketplace/redhat-operators-wp6hh" Nov 22 05:53:38 crc kubenswrapper[4818]: I1122 05:53:38.914665 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wp6hh" Nov 22 05:53:39 crc kubenswrapper[4818]: I1122 05:53:39.445078 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wp6hh"] Nov 22 05:53:40 crc kubenswrapper[4818]: I1122 05:53:40.128654 4818 generic.go:334] "Generic (PLEG): container finished" podID="ebcedfbe-da71-4e94-af5d-2fc3163fede9" containerID="931eb527ced41bf1a895d742189f1f4cbd2cf51bd00c1da5540bd8741470732d" exitCode=0 Nov 22 05:53:40 crc kubenswrapper[4818]: I1122 05:53:40.128714 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp6hh" event={"ID":"ebcedfbe-da71-4e94-af5d-2fc3163fede9","Type":"ContainerDied","Data":"931eb527ced41bf1a895d742189f1f4cbd2cf51bd00c1da5540bd8741470732d"} Nov 22 05:53:40 crc kubenswrapper[4818]: I1122 05:53:40.128944 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp6hh" event={"ID":"ebcedfbe-da71-4e94-af5d-2fc3163fede9","Type":"ContainerStarted","Data":"71510035dba8949d854e56bb557e254133b1694771fdd2a2dcd48d0e25f3fa98"} Nov 22 05:53:40 crc kubenswrapper[4818]: I1122 05:53:40.131364 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 05:53:41 crc kubenswrapper[4818]: I1122 05:53:41.140760 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp6hh" event={"ID":"ebcedfbe-da71-4e94-af5d-2fc3163fede9","Type":"ContainerStarted","Data":"41ec316948d555de52db83d5d8f8a206532db05345a12555b1943083b85c545b"} Nov 22 05:53:44 crc kubenswrapper[4818]: I1122 05:53:44.170918 4818 generic.go:334] "Generic (PLEG): container finished" podID="ebcedfbe-da71-4e94-af5d-2fc3163fede9" containerID="41ec316948d555de52db83d5d8f8a206532db05345a12555b1943083b85c545b" exitCode=0 Nov 22 05:53:44 crc kubenswrapper[4818]: I1122 05:53:44.171018 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp6hh" event={"ID":"ebcedfbe-da71-4e94-af5d-2fc3163fede9","Type":"ContainerDied","Data":"41ec316948d555de52db83d5d8f8a206532db05345a12555b1943083b85c545b"} Nov 22 05:53:45 crc kubenswrapper[4818]: I1122 05:53:45.181757 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp6hh" event={"ID":"ebcedfbe-da71-4e94-af5d-2fc3163fede9","Type":"ContainerStarted","Data":"f21375266dfce1c47af0fb7b1590e984daa2d107687da4507861e15eb79aeb94"} Nov 22 05:53:45 crc kubenswrapper[4818]: I1122 05:53:45.212523 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wp6hh" podStartSLOduration=2.7564117870000002 podStartE2EDuration="7.212495089s" podCreationTimestamp="2025-11-22 05:53:38 +0000 UTC" firstStartedPulling="2025-11-22 05:53:40.131121865 +0000 UTC m=+3972.705538392" lastFinishedPulling="2025-11-22 05:53:44.587205157 +0000 UTC m=+3977.161621694" observedRunningTime="2025-11-22 05:53:45.203506669 +0000 UTC m=+3977.777923236" watchObservedRunningTime="2025-11-22 05:53:45.212495089 +0000 UTC m=+3977.786911636" Nov 22 05:53:48 crc kubenswrapper[4818]: I1122 05:53:48.915194 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wp6hh" Nov 22 05:53:48 crc kubenswrapper[4818]: I1122 05:53:48.915724 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wp6hh" Nov 22 05:53:49 crc kubenswrapper[4818]: I1122 05:53:49.971916 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wp6hh" podUID="ebcedfbe-da71-4e94-af5d-2fc3163fede9" containerName="registry-server" probeResult="failure" output=< Nov 22 05:53:49 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Nov 22 05:53:49 crc kubenswrapper[4818]: > Nov 22 05:53:51 crc kubenswrapper[4818]: I1122 05:53:51.291395 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:53:51 crc kubenswrapper[4818]: E1122 05:53:51.291658 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:53:58 crc kubenswrapper[4818]: I1122 05:53:58.959990 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wp6hh" Nov 22 05:53:59 crc kubenswrapper[4818]: I1122 05:53:59.011543 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wp6hh" Nov 22 05:53:59 crc kubenswrapper[4818]: I1122 05:53:59.193431 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wp6hh"] Nov 22 05:54:00 crc kubenswrapper[4818]: I1122 05:54:00.328291 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wp6hh" podUID="ebcedfbe-da71-4e94-af5d-2fc3163fede9" containerName="registry-server" containerID="cri-o://f21375266dfce1c47af0fb7b1590e984daa2d107687da4507861e15eb79aeb94" gracePeriod=2 Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.176455 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wp6hh" Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.246551 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebcedfbe-da71-4e94-af5d-2fc3163fede9-utilities\") pod \"ebcedfbe-da71-4e94-af5d-2fc3163fede9\" (UID: \"ebcedfbe-da71-4e94-af5d-2fc3163fede9\") " Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.246698 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxfn9\" (UniqueName: \"kubernetes.io/projected/ebcedfbe-da71-4e94-af5d-2fc3163fede9-kube-api-access-zxfn9\") pod \"ebcedfbe-da71-4e94-af5d-2fc3163fede9\" (UID: \"ebcedfbe-da71-4e94-af5d-2fc3163fede9\") " Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.246739 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebcedfbe-da71-4e94-af5d-2fc3163fede9-catalog-content\") pod \"ebcedfbe-da71-4e94-af5d-2fc3163fede9\" (UID: \"ebcedfbe-da71-4e94-af5d-2fc3163fede9\") " Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.247414 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebcedfbe-da71-4e94-af5d-2fc3163fede9-utilities" (OuterVolumeSpecName: "utilities") pod "ebcedfbe-da71-4e94-af5d-2fc3163fede9" (UID: "ebcedfbe-da71-4e94-af5d-2fc3163fede9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.260581 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebcedfbe-da71-4e94-af5d-2fc3163fede9-kube-api-access-zxfn9" (OuterVolumeSpecName: "kube-api-access-zxfn9") pod "ebcedfbe-da71-4e94-af5d-2fc3163fede9" (UID: "ebcedfbe-da71-4e94-af5d-2fc3163fede9"). InnerVolumeSpecName "kube-api-access-zxfn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.338665 4818 generic.go:334] "Generic (PLEG): container finished" podID="ebcedfbe-da71-4e94-af5d-2fc3163fede9" containerID="f21375266dfce1c47af0fb7b1590e984daa2d107687da4507861e15eb79aeb94" exitCode=0 Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.338735 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wp6hh" Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.338733 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp6hh" event={"ID":"ebcedfbe-da71-4e94-af5d-2fc3163fede9","Type":"ContainerDied","Data":"f21375266dfce1c47af0fb7b1590e984daa2d107687da4507861e15eb79aeb94"} Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.338810 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp6hh" event={"ID":"ebcedfbe-da71-4e94-af5d-2fc3163fede9","Type":"ContainerDied","Data":"71510035dba8949d854e56bb557e254133b1694771fdd2a2dcd48d0e25f3fa98"} Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.338834 4818 scope.go:117] "RemoveContainer" containerID="f21375266dfce1c47af0fb7b1590e984daa2d107687da4507861e15eb79aeb94" Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.350655 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebcedfbe-da71-4e94-af5d-2fc3163fede9-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.350687 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxfn9\" (UniqueName: \"kubernetes.io/projected/ebcedfbe-da71-4e94-af5d-2fc3163fede9-kube-api-access-zxfn9\") on node \"crc\" DevicePath \"\"" Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.357726 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebcedfbe-da71-4e94-af5d-2fc3163fede9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ebcedfbe-da71-4e94-af5d-2fc3163fede9" (UID: "ebcedfbe-da71-4e94-af5d-2fc3163fede9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.359748 4818 scope.go:117] "RemoveContainer" containerID="41ec316948d555de52db83d5d8f8a206532db05345a12555b1943083b85c545b" Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.382773 4818 scope.go:117] "RemoveContainer" containerID="931eb527ced41bf1a895d742189f1f4cbd2cf51bd00c1da5540bd8741470732d" Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.437800 4818 scope.go:117] "RemoveContainer" containerID="f21375266dfce1c47af0fb7b1590e984daa2d107687da4507861e15eb79aeb94" Nov 22 05:54:01 crc kubenswrapper[4818]: E1122 05:54:01.438223 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f21375266dfce1c47af0fb7b1590e984daa2d107687da4507861e15eb79aeb94\": container with ID starting with f21375266dfce1c47af0fb7b1590e984daa2d107687da4507861e15eb79aeb94 not found: ID does not exist" containerID="f21375266dfce1c47af0fb7b1590e984daa2d107687da4507861e15eb79aeb94" Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.438281 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f21375266dfce1c47af0fb7b1590e984daa2d107687da4507861e15eb79aeb94"} err="failed to get container status \"f21375266dfce1c47af0fb7b1590e984daa2d107687da4507861e15eb79aeb94\": rpc error: code = NotFound desc = could not find container \"f21375266dfce1c47af0fb7b1590e984daa2d107687da4507861e15eb79aeb94\": container with ID starting with f21375266dfce1c47af0fb7b1590e984daa2d107687da4507861e15eb79aeb94 not found: ID does not exist" Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.438307 4818 scope.go:117] "RemoveContainer" containerID="41ec316948d555de52db83d5d8f8a206532db05345a12555b1943083b85c545b" Nov 22 05:54:01 crc kubenswrapper[4818]: E1122 05:54:01.438696 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41ec316948d555de52db83d5d8f8a206532db05345a12555b1943083b85c545b\": container with ID starting with 41ec316948d555de52db83d5d8f8a206532db05345a12555b1943083b85c545b not found: ID does not exist" containerID="41ec316948d555de52db83d5d8f8a206532db05345a12555b1943083b85c545b" Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.438725 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41ec316948d555de52db83d5d8f8a206532db05345a12555b1943083b85c545b"} err="failed to get container status \"41ec316948d555de52db83d5d8f8a206532db05345a12555b1943083b85c545b\": rpc error: code = NotFound desc = could not find container \"41ec316948d555de52db83d5d8f8a206532db05345a12555b1943083b85c545b\": container with ID starting with 41ec316948d555de52db83d5d8f8a206532db05345a12555b1943083b85c545b not found: ID does not exist" Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.438751 4818 scope.go:117] "RemoveContainer" containerID="931eb527ced41bf1a895d742189f1f4cbd2cf51bd00c1da5540bd8741470732d" Nov 22 05:54:01 crc kubenswrapper[4818]: E1122 05:54:01.439130 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"931eb527ced41bf1a895d742189f1f4cbd2cf51bd00c1da5540bd8741470732d\": container with ID starting with 931eb527ced41bf1a895d742189f1f4cbd2cf51bd00c1da5540bd8741470732d not found: ID does not exist" containerID="931eb527ced41bf1a895d742189f1f4cbd2cf51bd00c1da5540bd8741470732d" Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.439180 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"931eb527ced41bf1a895d742189f1f4cbd2cf51bd00c1da5540bd8741470732d"} err="failed to get container status \"931eb527ced41bf1a895d742189f1f4cbd2cf51bd00c1da5540bd8741470732d\": rpc error: code = NotFound desc = could not find container \"931eb527ced41bf1a895d742189f1f4cbd2cf51bd00c1da5540bd8741470732d\": container with ID starting with 931eb527ced41bf1a895d742189f1f4cbd2cf51bd00c1da5540bd8741470732d not found: ID does not exist" Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.452894 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebcedfbe-da71-4e94-af5d-2fc3163fede9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.671782 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wp6hh"] Nov 22 05:54:01 crc kubenswrapper[4818]: I1122 05:54:01.685087 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wp6hh"] Nov 22 05:54:02 crc kubenswrapper[4818]: I1122 05:54:02.290823 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:54:02 crc kubenswrapper[4818]: E1122 05:54:02.291314 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:54:02 crc kubenswrapper[4818]: I1122 05:54:02.301873 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebcedfbe-da71-4e94-af5d-2fc3163fede9" path="/var/lib/kubelet/pods/ebcedfbe-da71-4e94-af5d-2fc3163fede9/volumes" Nov 22 05:54:13 crc kubenswrapper[4818]: I1122 05:54:13.292149 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:54:13 crc kubenswrapper[4818]: E1122 05:54:13.293534 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:54:28 crc kubenswrapper[4818]: I1122 05:54:28.319760 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:54:28 crc kubenswrapper[4818]: E1122 05:54:28.320853 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:54:40 crc kubenswrapper[4818]: I1122 05:54:40.291636 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:54:40 crc kubenswrapper[4818]: E1122 05:54:40.292370 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 05:54:52 crc kubenswrapper[4818]: I1122 05:54:52.292410 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 05:54:52 crc kubenswrapper[4818]: I1122 05:54:52.839551 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"87e53488d73d6090d237f383bc00e58356c78446d810b06ded7d6522cfcda586"} Nov 22 05:55:32 crc kubenswrapper[4818]: I1122 05:55:32.040549 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-v2bv5"] Nov 22 05:55:32 crc kubenswrapper[4818]: I1122 05:55:32.052104 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-50ec-account-create-zq678"] Nov 22 05:55:32 crc kubenswrapper[4818]: I1122 05:55:32.061392 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-50ec-account-create-zq678"] Nov 22 05:55:32 crc kubenswrapper[4818]: I1122 05:55:32.069224 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-v2bv5"] Nov 22 05:55:32 crc kubenswrapper[4818]: I1122 05:55:32.301322 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ca6675b-db38-49ba-a7ab-396d5d6dc54f" path="/var/lib/kubelet/pods/1ca6675b-db38-49ba-a7ab-396d5d6dc54f/volumes" Nov 22 05:55:32 crc kubenswrapper[4818]: I1122 05:55:32.301869 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="384fe684-9694-4069-b37e-ad50f83ef867" path="/var/lib/kubelet/pods/384fe684-9694-4069-b37e-ad50f83ef867/volumes" Nov 22 05:55:44 crc kubenswrapper[4818]: I1122 05:55:44.098527 4818 scope.go:117] "RemoveContainer" containerID="a9f00296718de0e7ebca022ccc1ce4981fe7d40d35576e1abf7278830f38ceb8" Nov 22 05:55:44 crc kubenswrapper[4818]: I1122 05:55:44.135922 4818 scope.go:117] "RemoveContainer" containerID="1bc8c9b47327022b405fcdcb1246ffd08cae2061a2c32c8301a924e45630d081" Nov 22 05:55:56 crc kubenswrapper[4818]: I1122 05:55:56.114452 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vzw6z"] Nov 22 05:55:56 crc kubenswrapper[4818]: E1122 05:55:56.124676 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebcedfbe-da71-4e94-af5d-2fc3163fede9" containerName="extract-content" Nov 22 05:55:56 crc kubenswrapper[4818]: I1122 05:55:56.124700 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebcedfbe-da71-4e94-af5d-2fc3163fede9" containerName="extract-content" Nov 22 05:55:56 crc kubenswrapper[4818]: E1122 05:55:56.124731 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebcedfbe-da71-4e94-af5d-2fc3163fede9" containerName="registry-server" Nov 22 05:55:56 crc kubenswrapper[4818]: I1122 05:55:56.124739 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebcedfbe-da71-4e94-af5d-2fc3163fede9" containerName="registry-server" Nov 22 05:55:56 crc kubenswrapper[4818]: E1122 05:55:56.124774 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebcedfbe-da71-4e94-af5d-2fc3163fede9" containerName="extract-utilities" Nov 22 05:55:56 crc kubenswrapper[4818]: I1122 05:55:56.124780 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebcedfbe-da71-4e94-af5d-2fc3163fede9" containerName="extract-utilities" Nov 22 05:55:56 crc kubenswrapper[4818]: I1122 05:55:56.124961 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebcedfbe-da71-4e94-af5d-2fc3163fede9" containerName="registry-server" Nov 22 05:55:56 crc kubenswrapper[4818]: I1122 05:55:56.126328 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vzw6z" Nov 22 05:55:56 crc kubenswrapper[4818]: I1122 05:55:56.129069 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vzw6z"] Nov 22 05:55:56 crc kubenswrapper[4818]: I1122 05:55:56.247705 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01569daf-30e0-4f74-8e75-b7c3296f6ba3-utilities\") pod \"certified-operators-vzw6z\" (UID: \"01569daf-30e0-4f74-8e75-b7c3296f6ba3\") " pod="openshift-marketplace/certified-operators-vzw6z" Nov 22 05:55:56 crc kubenswrapper[4818]: I1122 05:55:56.247945 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01569daf-30e0-4f74-8e75-b7c3296f6ba3-catalog-content\") pod \"certified-operators-vzw6z\" (UID: \"01569daf-30e0-4f74-8e75-b7c3296f6ba3\") " pod="openshift-marketplace/certified-operators-vzw6z" Nov 22 05:55:56 crc kubenswrapper[4818]: I1122 05:55:56.247968 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bm5tp\" (UniqueName: \"kubernetes.io/projected/01569daf-30e0-4f74-8e75-b7c3296f6ba3-kube-api-access-bm5tp\") pod \"certified-operators-vzw6z\" (UID: \"01569daf-30e0-4f74-8e75-b7c3296f6ba3\") " pod="openshift-marketplace/certified-operators-vzw6z" Nov 22 05:55:56 crc kubenswrapper[4818]: I1122 05:55:56.350290 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01569daf-30e0-4f74-8e75-b7c3296f6ba3-catalog-content\") pod \"certified-operators-vzw6z\" (UID: \"01569daf-30e0-4f74-8e75-b7c3296f6ba3\") " pod="openshift-marketplace/certified-operators-vzw6z" Nov 22 05:55:56 crc kubenswrapper[4818]: I1122 05:55:56.350337 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm5tp\" (UniqueName: \"kubernetes.io/projected/01569daf-30e0-4f74-8e75-b7c3296f6ba3-kube-api-access-bm5tp\") pod \"certified-operators-vzw6z\" (UID: \"01569daf-30e0-4f74-8e75-b7c3296f6ba3\") " pod="openshift-marketplace/certified-operators-vzw6z" Nov 22 05:55:56 crc kubenswrapper[4818]: I1122 05:55:56.350451 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01569daf-30e0-4f74-8e75-b7c3296f6ba3-utilities\") pod \"certified-operators-vzw6z\" (UID: \"01569daf-30e0-4f74-8e75-b7c3296f6ba3\") " pod="openshift-marketplace/certified-operators-vzw6z" Nov 22 05:55:56 crc kubenswrapper[4818]: I1122 05:55:56.350960 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01569daf-30e0-4f74-8e75-b7c3296f6ba3-utilities\") pod \"certified-operators-vzw6z\" (UID: \"01569daf-30e0-4f74-8e75-b7c3296f6ba3\") " pod="openshift-marketplace/certified-operators-vzw6z" Nov 22 05:55:56 crc kubenswrapper[4818]: I1122 05:55:56.351196 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01569daf-30e0-4f74-8e75-b7c3296f6ba3-catalog-content\") pod \"certified-operators-vzw6z\" (UID: \"01569daf-30e0-4f74-8e75-b7c3296f6ba3\") " pod="openshift-marketplace/certified-operators-vzw6z" Nov 22 05:55:56 crc kubenswrapper[4818]: I1122 05:55:56.372724 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bm5tp\" (UniqueName: \"kubernetes.io/projected/01569daf-30e0-4f74-8e75-b7c3296f6ba3-kube-api-access-bm5tp\") pod \"certified-operators-vzw6z\" (UID: \"01569daf-30e0-4f74-8e75-b7c3296f6ba3\") " pod="openshift-marketplace/certified-operators-vzw6z" Nov 22 05:55:56 crc kubenswrapper[4818]: I1122 05:55:56.447833 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vzw6z" Nov 22 05:55:57 crc kubenswrapper[4818]: I1122 05:55:57.012744 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vzw6z"] Nov 22 05:55:57 crc kubenswrapper[4818]: I1122 05:55:57.459518 4818 generic.go:334] "Generic (PLEG): container finished" podID="01569daf-30e0-4f74-8e75-b7c3296f6ba3" containerID="b82d3048dd2e1d80365f90e47c295fe86bd27c0ca75181a1267d90b2e0f2318d" exitCode=0 Nov 22 05:55:57 crc kubenswrapper[4818]: I1122 05:55:57.459746 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzw6z" event={"ID":"01569daf-30e0-4f74-8e75-b7c3296f6ba3","Type":"ContainerDied","Data":"b82d3048dd2e1d80365f90e47c295fe86bd27c0ca75181a1267d90b2e0f2318d"} Nov 22 05:55:57 crc kubenswrapper[4818]: I1122 05:55:57.459859 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzw6z" event={"ID":"01569daf-30e0-4f74-8e75-b7c3296f6ba3","Type":"ContainerStarted","Data":"b8139244ae916dca8b645312a5bf013d303278def7e2fcd34164c7a867d44c4b"} Nov 22 05:55:58 crc kubenswrapper[4818]: I1122 05:55:58.471654 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzw6z" event={"ID":"01569daf-30e0-4f74-8e75-b7c3296f6ba3","Type":"ContainerStarted","Data":"5e768987fd283665d39fa5c72baac8246c13d4185f77c227b2ad6495a9889039"} Nov 22 05:55:59 crc kubenswrapper[4818]: I1122 05:55:59.482624 4818 generic.go:334] "Generic (PLEG): container finished" podID="01569daf-30e0-4f74-8e75-b7c3296f6ba3" containerID="5e768987fd283665d39fa5c72baac8246c13d4185f77c227b2ad6495a9889039" exitCode=0 Nov 22 05:55:59 crc kubenswrapper[4818]: I1122 05:55:59.482830 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzw6z" event={"ID":"01569daf-30e0-4f74-8e75-b7c3296f6ba3","Type":"ContainerDied","Data":"5e768987fd283665d39fa5c72baac8246c13d4185f77c227b2ad6495a9889039"} Nov 22 05:56:00 crc kubenswrapper[4818]: I1122 05:56:00.051856 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-2xttz"] Nov 22 05:56:00 crc kubenswrapper[4818]: I1122 05:56:00.061766 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-2xttz"] Nov 22 05:56:00 crc kubenswrapper[4818]: I1122 05:56:00.306096 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64f4e98d-2953-472c-852a-424b72c513a9" path="/var/lib/kubelet/pods/64f4e98d-2953-472c-852a-424b72c513a9/volumes" Nov 22 05:56:00 crc kubenswrapper[4818]: I1122 05:56:00.502789 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzw6z" event={"ID":"01569daf-30e0-4f74-8e75-b7c3296f6ba3","Type":"ContainerStarted","Data":"97affd775bfcd1d3152ee6bcdf244bbad0c120b837f765462998d7de0c4986b2"} Nov 22 05:56:00 crc kubenswrapper[4818]: I1122 05:56:00.521367 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vzw6z" podStartSLOduration=1.948345898 podStartE2EDuration="4.521353562s" podCreationTimestamp="2025-11-22 05:55:56 +0000 UTC" firstStartedPulling="2025-11-22 05:55:57.461149655 +0000 UTC m=+4110.035566192" lastFinishedPulling="2025-11-22 05:56:00.034157329 +0000 UTC m=+4112.608573856" observedRunningTime="2025-11-22 05:56:00.519335359 +0000 UTC m=+4113.093751896" watchObservedRunningTime="2025-11-22 05:56:00.521353562 +0000 UTC m=+4113.095770089" Nov 22 05:56:06 crc kubenswrapper[4818]: I1122 05:56:06.448236 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vzw6z" Nov 22 05:56:06 crc kubenswrapper[4818]: I1122 05:56:06.448833 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vzw6z" Nov 22 05:56:06 crc kubenswrapper[4818]: I1122 05:56:06.500210 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vzw6z" Nov 22 05:56:06 crc kubenswrapper[4818]: I1122 05:56:06.605021 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vzw6z" Nov 22 05:56:06 crc kubenswrapper[4818]: I1122 05:56:06.735386 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vzw6z"] Nov 22 05:56:08 crc kubenswrapper[4818]: I1122 05:56:08.573533 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vzw6z" podUID="01569daf-30e0-4f74-8e75-b7c3296f6ba3" containerName="registry-server" containerID="cri-o://97affd775bfcd1d3152ee6bcdf244bbad0c120b837f765462998d7de0c4986b2" gracePeriod=2 Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.158582 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vzw6z" Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.265162 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bm5tp\" (UniqueName: \"kubernetes.io/projected/01569daf-30e0-4f74-8e75-b7c3296f6ba3-kube-api-access-bm5tp\") pod \"01569daf-30e0-4f74-8e75-b7c3296f6ba3\" (UID: \"01569daf-30e0-4f74-8e75-b7c3296f6ba3\") " Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.265350 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01569daf-30e0-4f74-8e75-b7c3296f6ba3-utilities\") pod \"01569daf-30e0-4f74-8e75-b7c3296f6ba3\" (UID: \"01569daf-30e0-4f74-8e75-b7c3296f6ba3\") " Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.265404 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01569daf-30e0-4f74-8e75-b7c3296f6ba3-catalog-content\") pod \"01569daf-30e0-4f74-8e75-b7c3296f6ba3\" (UID: \"01569daf-30e0-4f74-8e75-b7c3296f6ba3\") " Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.266384 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01569daf-30e0-4f74-8e75-b7c3296f6ba3-utilities" (OuterVolumeSpecName: "utilities") pod "01569daf-30e0-4f74-8e75-b7c3296f6ba3" (UID: "01569daf-30e0-4f74-8e75-b7c3296f6ba3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.270755 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01569daf-30e0-4f74-8e75-b7c3296f6ba3-kube-api-access-bm5tp" (OuterVolumeSpecName: "kube-api-access-bm5tp") pod "01569daf-30e0-4f74-8e75-b7c3296f6ba3" (UID: "01569daf-30e0-4f74-8e75-b7c3296f6ba3"). InnerVolumeSpecName "kube-api-access-bm5tp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.368153 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bm5tp\" (UniqueName: \"kubernetes.io/projected/01569daf-30e0-4f74-8e75-b7c3296f6ba3-kube-api-access-bm5tp\") on node \"crc\" DevicePath \"\"" Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.368192 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01569daf-30e0-4f74-8e75-b7c3296f6ba3-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.584836 4818 generic.go:334] "Generic (PLEG): container finished" podID="01569daf-30e0-4f74-8e75-b7c3296f6ba3" containerID="97affd775bfcd1d3152ee6bcdf244bbad0c120b837f765462998d7de0c4986b2" exitCode=0 Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.584892 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzw6z" event={"ID":"01569daf-30e0-4f74-8e75-b7c3296f6ba3","Type":"ContainerDied","Data":"97affd775bfcd1d3152ee6bcdf244bbad0c120b837f765462998d7de0c4986b2"} Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.584931 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzw6z" event={"ID":"01569daf-30e0-4f74-8e75-b7c3296f6ba3","Type":"ContainerDied","Data":"b8139244ae916dca8b645312a5bf013d303278def7e2fcd34164c7a867d44c4b"} Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.584960 4818 scope.go:117] "RemoveContainer" containerID="97affd775bfcd1d3152ee6bcdf244bbad0c120b837f765462998d7de0c4986b2" Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.584896 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vzw6z" Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.608982 4818 scope.go:117] "RemoveContainer" containerID="5e768987fd283665d39fa5c72baac8246c13d4185f77c227b2ad6495a9889039" Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.642590 4818 scope.go:117] "RemoveContainer" containerID="b82d3048dd2e1d80365f90e47c295fe86bd27c0ca75181a1267d90b2e0f2318d" Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.692027 4818 scope.go:117] "RemoveContainer" containerID="97affd775bfcd1d3152ee6bcdf244bbad0c120b837f765462998d7de0c4986b2" Nov 22 05:56:09 crc kubenswrapper[4818]: E1122 05:56:09.692958 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97affd775bfcd1d3152ee6bcdf244bbad0c120b837f765462998d7de0c4986b2\": container with ID starting with 97affd775bfcd1d3152ee6bcdf244bbad0c120b837f765462998d7de0c4986b2 not found: ID does not exist" containerID="97affd775bfcd1d3152ee6bcdf244bbad0c120b837f765462998d7de0c4986b2" Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.693015 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97affd775bfcd1d3152ee6bcdf244bbad0c120b837f765462998d7de0c4986b2"} err="failed to get container status \"97affd775bfcd1d3152ee6bcdf244bbad0c120b837f765462998d7de0c4986b2\": rpc error: code = NotFound desc = could not find container \"97affd775bfcd1d3152ee6bcdf244bbad0c120b837f765462998d7de0c4986b2\": container with ID starting with 97affd775bfcd1d3152ee6bcdf244bbad0c120b837f765462998d7de0c4986b2 not found: ID does not exist" Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.693043 4818 scope.go:117] "RemoveContainer" containerID="5e768987fd283665d39fa5c72baac8246c13d4185f77c227b2ad6495a9889039" Nov 22 05:56:09 crc kubenswrapper[4818]: E1122 05:56:09.693458 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e768987fd283665d39fa5c72baac8246c13d4185f77c227b2ad6495a9889039\": container with ID starting with 5e768987fd283665d39fa5c72baac8246c13d4185f77c227b2ad6495a9889039 not found: ID does not exist" containerID="5e768987fd283665d39fa5c72baac8246c13d4185f77c227b2ad6495a9889039" Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.693489 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e768987fd283665d39fa5c72baac8246c13d4185f77c227b2ad6495a9889039"} err="failed to get container status \"5e768987fd283665d39fa5c72baac8246c13d4185f77c227b2ad6495a9889039\": rpc error: code = NotFound desc = could not find container \"5e768987fd283665d39fa5c72baac8246c13d4185f77c227b2ad6495a9889039\": container with ID starting with 5e768987fd283665d39fa5c72baac8246c13d4185f77c227b2ad6495a9889039 not found: ID does not exist" Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.693518 4818 scope.go:117] "RemoveContainer" containerID="b82d3048dd2e1d80365f90e47c295fe86bd27c0ca75181a1267d90b2e0f2318d" Nov 22 05:56:09 crc kubenswrapper[4818]: E1122 05:56:09.693942 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b82d3048dd2e1d80365f90e47c295fe86bd27c0ca75181a1267d90b2e0f2318d\": container with ID starting with b82d3048dd2e1d80365f90e47c295fe86bd27c0ca75181a1267d90b2e0f2318d not found: ID does not exist" containerID="b82d3048dd2e1d80365f90e47c295fe86bd27c0ca75181a1267d90b2e0f2318d" Nov 22 05:56:09 crc kubenswrapper[4818]: I1122 05:56:09.694005 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b82d3048dd2e1d80365f90e47c295fe86bd27c0ca75181a1267d90b2e0f2318d"} err="failed to get container status \"b82d3048dd2e1d80365f90e47c295fe86bd27c0ca75181a1267d90b2e0f2318d\": rpc error: code = NotFound desc = could not find container \"b82d3048dd2e1d80365f90e47c295fe86bd27c0ca75181a1267d90b2e0f2318d\": container with ID starting with b82d3048dd2e1d80365f90e47c295fe86bd27c0ca75181a1267d90b2e0f2318d not found: ID does not exist" Nov 22 05:56:10 crc kubenswrapper[4818]: I1122 05:56:10.588147 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01569daf-30e0-4f74-8e75-b7c3296f6ba3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "01569daf-30e0-4f74-8e75-b7c3296f6ba3" (UID: "01569daf-30e0-4f74-8e75-b7c3296f6ba3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:56:10 crc kubenswrapper[4818]: I1122 05:56:10.610752 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01569daf-30e0-4f74-8e75-b7c3296f6ba3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 05:56:10 crc kubenswrapper[4818]: I1122 05:56:10.867193 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vzw6z"] Nov 22 05:56:10 crc kubenswrapper[4818]: I1122 05:56:10.887015 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vzw6z"] Nov 22 05:56:10 crc kubenswrapper[4818]: E1122 05:56:10.939006 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01569daf_30e0_4f74_8e75_b7c3296f6ba3.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01569daf_30e0_4f74_8e75_b7c3296f6ba3.slice/crio-b8139244ae916dca8b645312a5bf013d303278def7e2fcd34164c7a867d44c4b\": RecentStats: unable to find data in memory cache]" Nov 22 05:56:12 crc kubenswrapper[4818]: I1122 05:56:12.307097 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01569daf-30e0-4f74-8e75-b7c3296f6ba3" path="/var/lib/kubelet/pods/01569daf-30e0-4f74-8e75-b7c3296f6ba3/volumes" Nov 22 05:56:19 crc kubenswrapper[4818]: I1122 05:56:19.249183 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cdjr2"] Nov 22 05:56:19 crc kubenswrapper[4818]: E1122 05:56:19.250024 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01569daf-30e0-4f74-8e75-b7c3296f6ba3" containerName="extract-content" Nov 22 05:56:19 crc kubenswrapper[4818]: I1122 05:56:19.250036 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="01569daf-30e0-4f74-8e75-b7c3296f6ba3" containerName="extract-content" Nov 22 05:56:19 crc kubenswrapper[4818]: E1122 05:56:19.250075 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01569daf-30e0-4f74-8e75-b7c3296f6ba3" containerName="extract-utilities" Nov 22 05:56:19 crc kubenswrapper[4818]: I1122 05:56:19.250083 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="01569daf-30e0-4f74-8e75-b7c3296f6ba3" containerName="extract-utilities" Nov 22 05:56:19 crc kubenswrapper[4818]: E1122 05:56:19.250092 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01569daf-30e0-4f74-8e75-b7c3296f6ba3" containerName="registry-server" Nov 22 05:56:19 crc kubenswrapper[4818]: I1122 05:56:19.250099 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="01569daf-30e0-4f74-8e75-b7c3296f6ba3" containerName="registry-server" Nov 22 05:56:19 crc kubenswrapper[4818]: I1122 05:56:19.250302 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="01569daf-30e0-4f74-8e75-b7c3296f6ba3" containerName="registry-server" Nov 22 05:56:19 crc kubenswrapper[4818]: I1122 05:56:19.251556 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cdjr2" Nov 22 05:56:19 crc kubenswrapper[4818]: I1122 05:56:19.267168 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cdjr2"] Nov 22 05:56:19 crc kubenswrapper[4818]: I1122 05:56:19.285621 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd603ec5-ddaf-4d51-a4fc-09442ae6692e-utilities\") pod \"community-operators-cdjr2\" (UID: \"fd603ec5-ddaf-4d51-a4fc-09442ae6692e\") " pod="openshift-marketplace/community-operators-cdjr2" Nov 22 05:56:19 crc kubenswrapper[4818]: I1122 05:56:19.286007 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd603ec5-ddaf-4d51-a4fc-09442ae6692e-catalog-content\") pod \"community-operators-cdjr2\" (UID: \"fd603ec5-ddaf-4d51-a4fc-09442ae6692e\") " pod="openshift-marketplace/community-operators-cdjr2" Nov 22 05:56:19 crc kubenswrapper[4818]: I1122 05:56:19.286152 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhx4j\" (UniqueName: \"kubernetes.io/projected/fd603ec5-ddaf-4d51-a4fc-09442ae6692e-kube-api-access-lhx4j\") pod \"community-operators-cdjr2\" (UID: \"fd603ec5-ddaf-4d51-a4fc-09442ae6692e\") " pod="openshift-marketplace/community-operators-cdjr2" Nov 22 05:56:19 crc kubenswrapper[4818]: I1122 05:56:19.387925 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd603ec5-ddaf-4d51-a4fc-09442ae6692e-catalog-content\") pod \"community-operators-cdjr2\" (UID: \"fd603ec5-ddaf-4d51-a4fc-09442ae6692e\") " pod="openshift-marketplace/community-operators-cdjr2" Nov 22 05:56:19 crc kubenswrapper[4818]: I1122 05:56:19.388042 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhx4j\" (UniqueName: \"kubernetes.io/projected/fd603ec5-ddaf-4d51-a4fc-09442ae6692e-kube-api-access-lhx4j\") pod \"community-operators-cdjr2\" (UID: \"fd603ec5-ddaf-4d51-a4fc-09442ae6692e\") " pod="openshift-marketplace/community-operators-cdjr2" Nov 22 05:56:19 crc kubenswrapper[4818]: I1122 05:56:19.388226 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd603ec5-ddaf-4d51-a4fc-09442ae6692e-utilities\") pod \"community-operators-cdjr2\" (UID: \"fd603ec5-ddaf-4d51-a4fc-09442ae6692e\") " pod="openshift-marketplace/community-operators-cdjr2" Nov 22 05:56:19 crc kubenswrapper[4818]: I1122 05:56:19.389919 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd603ec5-ddaf-4d51-a4fc-09442ae6692e-catalog-content\") pod \"community-operators-cdjr2\" (UID: \"fd603ec5-ddaf-4d51-a4fc-09442ae6692e\") " pod="openshift-marketplace/community-operators-cdjr2" Nov 22 05:56:19 crc kubenswrapper[4818]: I1122 05:56:19.390846 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd603ec5-ddaf-4d51-a4fc-09442ae6692e-utilities\") pod \"community-operators-cdjr2\" (UID: \"fd603ec5-ddaf-4d51-a4fc-09442ae6692e\") " pod="openshift-marketplace/community-operators-cdjr2" Nov 22 05:56:19 crc kubenswrapper[4818]: I1122 05:56:19.414176 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhx4j\" (UniqueName: \"kubernetes.io/projected/fd603ec5-ddaf-4d51-a4fc-09442ae6692e-kube-api-access-lhx4j\") pod \"community-operators-cdjr2\" (UID: \"fd603ec5-ddaf-4d51-a4fc-09442ae6692e\") " pod="openshift-marketplace/community-operators-cdjr2" Nov 22 05:56:19 crc kubenswrapper[4818]: I1122 05:56:19.572034 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cdjr2" Nov 22 05:56:20 crc kubenswrapper[4818]: I1122 05:56:20.091992 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cdjr2"] Nov 22 05:56:20 crc kubenswrapper[4818]: I1122 05:56:20.680531 4818 generic.go:334] "Generic (PLEG): container finished" podID="fd603ec5-ddaf-4d51-a4fc-09442ae6692e" containerID="d73801f47ac0d4e60b9f42cea49765fd3626ef3fedea19f1fdbdeaf8a96b3a7c" exitCode=0 Nov 22 05:56:20 crc kubenswrapper[4818]: I1122 05:56:20.680605 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdjr2" event={"ID":"fd603ec5-ddaf-4d51-a4fc-09442ae6692e","Type":"ContainerDied","Data":"d73801f47ac0d4e60b9f42cea49765fd3626ef3fedea19f1fdbdeaf8a96b3a7c"} Nov 22 05:56:20 crc kubenswrapper[4818]: I1122 05:56:20.681703 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdjr2" event={"ID":"fd603ec5-ddaf-4d51-a4fc-09442ae6692e","Type":"ContainerStarted","Data":"96fe9321d20607f0bb7140004e354ce1b1b61eeb73f3da9e0349c25affd30061"} Nov 22 05:56:23 crc kubenswrapper[4818]: I1122 05:56:23.065362 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lwcql"] Nov 22 05:56:23 crc kubenswrapper[4818]: I1122 05:56:23.068034 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lwcql" Nov 22 05:56:23 crc kubenswrapper[4818]: I1122 05:56:23.075811 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lwcql"] Nov 22 05:56:23 crc kubenswrapper[4818]: I1122 05:56:23.078311 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72a97bea-43fc-40b7-abd0-854197951ac8-catalog-content\") pod \"redhat-marketplace-lwcql\" (UID: \"72a97bea-43fc-40b7-abd0-854197951ac8\") " pod="openshift-marketplace/redhat-marketplace-lwcql" Nov 22 05:56:23 crc kubenswrapper[4818]: I1122 05:56:23.078583 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwxwm\" (UniqueName: \"kubernetes.io/projected/72a97bea-43fc-40b7-abd0-854197951ac8-kube-api-access-wwxwm\") pod \"redhat-marketplace-lwcql\" (UID: \"72a97bea-43fc-40b7-abd0-854197951ac8\") " pod="openshift-marketplace/redhat-marketplace-lwcql" Nov 22 05:56:23 crc kubenswrapper[4818]: I1122 05:56:23.078631 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72a97bea-43fc-40b7-abd0-854197951ac8-utilities\") pod \"redhat-marketplace-lwcql\" (UID: \"72a97bea-43fc-40b7-abd0-854197951ac8\") " pod="openshift-marketplace/redhat-marketplace-lwcql" Nov 22 05:56:23 crc kubenswrapper[4818]: I1122 05:56:23.180827 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwxwm\" (UniqueName: \"kubernetes.io/projected/72a97bea-43fc-40b7-abd0-854197951ac8-kube-api-access-wwxwm\") pod \"redhat-marketplace-lwcql\" (UID: \"72a97bea-43fc-40b7-abd0-854197951ac8\") " pod="openshift-marketplace/redhat-marketplace-lwcql" Nov 22 05:56:23 crc kubenswrapper[4818]: I1122 05:56:23.180884 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72a97bea-43fc-40b7-abd0-854197951ac8-utilities\") pod \"redhat-marketplace-lwcql\" (UID: \"72a97bea-43fc-40b7-abd0-854197951ac8\") " pod="openshift-marketplace/redhat-marketplace-lwcql" Nov 22 05:56:23 crc kubenswrapper[4818]: I1122 05:56:23.180931 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72a97bea-43fc-40b7-abd0-854197951ac8-catalog-content\") pod \"redhat-marketplace-lwcql\" (UID: \"72a97bea-43fc-40b7-abd0-854197951ac8\") " pod="openshift-marketplace/redhat-marketplace-lwcql" Nov 22 05:56:23 crc kubenswrapper[4818]: I1122 05:56:23.181429 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72a97bea-43fc-40b7-abd0-854197951ac8-utilities\") pod \"redhat-marketplace-lwcql\" (UID: \"72a97bea-43fc-40b7-abd0-854197951ac8\") " pod="openshift-marketplace/redhat-marketplace-lwcql" Nov 22 05:56:23 crc kubenswrapper[4818]: I1122 05:56:23.181452 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72a97bea-43fc-40b7-abd0-854197951ac8-catalog-content\") pod \"redhat-marketplace-lwcql\" (UID: \"72a97bea-43fc-40b7-abd0-854197951ac8\") " pod="openshift-marketplace/redhat-marketplace-lwcql" Nov 22 05:56:23 crc kubenswrapper[4818]: I1122 05:56:23.201590 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwxwm\" (UniqueName: \"kubernetes.io/projected/72a97bea-43fc-40b7-abd0-854197951ac8-kube-api-access-wwxwm\") pod \"redhat-marketplace-lwcql\" (UID: \"72a97bea-43fc-40b7-abd0-854197951ac8\") " pod="openshift-marketplace/redhat-marketplace-lwcql" Nov 22 05:56:23 crc kubenswrapper[4818]: I1122 05:56:23.384103 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lwcql" Nov 22 05:56:27 crc kubenswrapper[4818]: I1122 05:56:27.659028 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lwcql"] Nov 22 05:56:27 crc kubenswrapper[4818]: I1122 05:56:27.747613 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lwcql" event={"ID":"72a97bea-43fc-40b7-abd0-854197951ac8","Type":"ContainerStarted","Data":"2ba3d5e7628203146a6a6dc0d7248963636093a339195a52bab879140b4725c0"} Nov 22 05:56:28 crc kubenswrapper[4818]: I1122 05:56:28.760903 4818 generic.go:334] "Generic (PLEG): container finished" podID="fd603ec5-ddaf-4d51-a4fc-09442ae6692e" containerID="3600af030a0e2e6cef77056d9d0e4e513570005f5595550436f240c5ab8f9562" exitCode=0 Nov 22 05:56:28 crc kubenswrapper[4818]: I1122 05:56:28.760971 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdjr2" event={"ID":"fd603ec5-ddaf-4d51-a4fc-09442ae6692e","Type":"ContainerDied","Data":"3600af030a0e2e6cef77056d9d0e4e513570005f5595550436f240c5ab8f9562"} Nov 22 05:56:28 crc kubenswrapper[4818]: I1122 05:56:28.763243 4818 generic.go:334] "Generic (PLEG): container finished" podID="72a97bea-43fc-40b7-abd0-854197951ac8" containerID="c7273c8003724742709026453738c67574159c11278e521bdc50e80bc06d819d" exitCode=0 Nov 22 05:56:28 crc kubenswrapper[4818]: I1122 05:56:28.763294 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lwcql" event={"ID":"72a97bea-43fc-40b7-abd0-854197951ac8","Type":"ContainerDied","Data":"c7273c8003724742709026453738c67574159c11278e521bdc50e80bc06d819d"} Nov 22 05:56:30 crc kubenswrapper[4818]: I1122 05:56:30.785672 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdjr2" event={"ID":"fd603ec5-ddaf-4d51-a4fc-09442ae6692e","Type":"ContainerStarted","Data":"8708eb9bcaacc007c881fc87290a6e2c3e2977db6625f6814620c55d907dad7b"} Nov 22 05:56:30 crc kubenswrapper[4818]: I1122 05:56:30.788847 4818 generic.go:334] "Generic (PLEG): container finished" podID="72a97bea-43fc-40b7-abd0-854197951ac8" containerID="6ea72e9bfee027b7f049dd5edb02d3f848f70aecda4c13938a28e3a50efaf00b" exitCode=0 Nov 22 05:56:30 crc kubenswrapper[4818]: I1122 05:56:30.788903 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lwcql" event={"ID":"72a97bea-43fc-40b7-abd0-854197951ac8","Type":"ContainerDied","Data":"6ea72e9bfee027b7f049dd5edb02d3f848f70aecda4c13938a28e3a50efaf00b"} Nov 22 05:56:30 crc kubenswrapper[4818]: I1122 05:56:30.806167 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cdjr2" podStartSLOduration=2.303884263 podStartE2EDuration="11.806145528s" podCreationTimestamp="2025-11-22 05:56:19 +0000 UTC" firstStartedPulling="2025-11-22 05:56:20.682541977 +0000 UTC m=+4133.256958494" lastFinishedPulling="2025-11-22 05:56:30.184803232 +0000 UTC m=+4142.759219759" observedRunningTime="2025-11-22 05:56:30.804215197 +0000 UTC m=+4143.378631724" watchObservedRunningTime="2025-11-22 05:56:30.806145528 +0000 UTC m=+4143.380562065" Nov 22 05:56:31 crc kubenswrapper[4818]: I1122 05:56:31.799830 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lwcql" event={"ID":"72a97bea-43fc-40b7-abd0-854197951ac8","Type":"ContainerStarted","Data":"f3a4d787a2bf0a4ec2a1594251294a8a5472566da9704fba73961798116a4921"} Nov 22 05:56:31 crc kubenswrapper[4818]: I1122 05:56:31.823220 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lwcql" podStartSLOduration=6.287855105 podStartE2EDuration="8.823201844s" podCreationTimestamp="2025-11-22 05:56:23 +0000 UTC" firstStartedPulling="2025-11-22 05:56:28.764618229 +0000 UTC m=+4141.339034756" lastFinishedPulling="2025-11-22 05:56:31.299964968 +0000 UTC m=+4143.874381495" observedRunningTime="2025-11-22 05:56:31.815461707 +0000 UTC m=+4144.389878234" watchObservedRunningTime="2025-11-22 05:56:31.823201844 +0000 UTC m=+4144.397618371" Nov 22 05:56:33 crc kubenswrapper[4818]: I1122 05:56:33.384391 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lwcql" Nov 22 05:56:33 crc kubenswrapper[4818]: I1122 05:56:33.384766 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lwcql" Nov 22 05:56:33 crc kubenswrapper[4818]: I1122 05:56:33.464133 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lwcql" Nov 22 05:56:39 crc kubenswrapper[4818]: I1122 05:56:39.572204 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cdjr2" Nov 22 05:56:39 crc kubenswrapper[4818]: I1122 05:56:39.572770 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cdjr2" Nov 22 05:56:39 crc kubenswrapper[4818]: I1122 05:56:39.656945 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cdjr2" Nov 22 05:56:39 crc kubenswrapper[4818]: I1122 05:56:39.996973 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cdjr2" Nov 22 05:56:40 crc kubenswrapper[4818]: I1122 05:56:40.088919 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cdjr2"] Nov 22 05:56:40 crc kubenswrapper[4818]: I1122 05:56:40.146091 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k4j4l"] Nov 22 05:56:40 crc kubenswrapper[4818]: I1122 05:56:40.146405 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k4j4l" podUID="614051f2-a9cf-4084-9eb0-d5333b5b1d0e" containerName="registry-server" containerID="cri-o://a4886460724ba5684813c21db17c3989575dea69dff6dcd1076005977210e193" gracePeriod=2 Nov 22 05:56:40 crc kubenswrapper[4818]: I1122 05:56:40.639842 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k4j4l" Nov 22 05:56:40 crc kubenswrapper[4818]: I1122 05:56:40.812361 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/614051f2-a9cf-4084-9eb0-d5333b5b1d0e-utilities\") pod \"614051f2-a9cf-4084-9eb0-d5333b5b1d0e\" (UID: \"614051f2-a9cf-4084-9eb0-d5333b5b1d0e\") " Nov 22 05:56:40 crc kubenswrapper[4818]: I1122 05:56:40.812581 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/614051f2-a9cf-4084-9eb0-d5333b5b1d0e-catalog-content\") pod \"614051f2-a9cf-4084-9eb0-d5333b5b1d0e\" (UID: \"614051f2-a9cf-4084-9eb0-d5333b5b1d0e\") " Nov 22 05:56:40 crc kubenswrapper[4818]: I1122 05:56:40.812656 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bggh\" (UniqueName: \"kubernetes.io/projected/614051f2-a9cf-4084-9eb0-d5333b5b1d0e-kube-api-access-8bggh\") pod \"614051f2-a9cf-4084-9eb0-d5333b5b1d0e\" (UID: \"614051f2-a9cf-4084-9eb0-d5333b5b1d0e\") " Nov 22 05:56:40 crc kubenswrapper[4818]: I1122 05:56:40.820575 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/614051f2-a9cf-4084-9eb0-d5333b5b1d0e-utilities" (OuterVolumeSpecName: "utilities") pod "614051f2-a9cf-4084-9eb0-d5333b5b1d0e" (UID: "614051f2-a9cf-4084-9eb0-d5333b5b1d0e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:56:40 crc kubenswrapper[4818]: I1122 05:56:40.825145 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/614051f2-a9cf-4084-9eb0-d5333b5b1d0e-kube-api-access-8bggh" (OuterVolumeSpecName: "kube-api-access-8bggh") pod "614051f2-a9cf-4084-9eb0-d5333b5b1d0e" (UID: "614051f2-a9cf-4084-9eb0-d5333b5b1d0e"). InnerVolumeSpecName "kube-api-access-8bggh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:56:40 crc kubenswrapper[4818]: I1122 05:56:40.915100 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/614051f2-a9cf-4084-9eb0-d5333b5b1d0e-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 05:56:40 crc kubenswrapper[4818]: I1122 05:56:40.915128 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bggh\" (UniqueName: \"kubernetes.io/projected/614051f2-a9cf-4084-9eb0-d5333b5b1d0e-kube-api-access-8bggh\") on node \"crc\" DevicePath \"\"" Nov 22 05:56:40 crc kubenswrapper[4818]: I1122 05:56:40.920144 4818 generic.go:334] "Generic (PLEG): container finished" podID="614051f2-a9cf-4084-9eb0-d5333b5b1d0e" containerID="a4886460724ba5684813c21db17c3989575dea69dff6dcd1076005977210e193" exitCode=0 Nov 22 05:56:40 crc kubenswrapper[4818]: I1122 05:56:40.920198 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k4j4l" Nov 22 05:56:40 crc kubenswrapper[4818]: I1122 05:56:40.920270 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4j4l" event={"ID":"614051f2-a9cf-4084-9eb0-d5333b5b1d0e","Type":"ContainerDied","Data":"a4886460724ba5684813c21db17c3989575dea69dff6dcd1076005977210e193"} Nov 22 05:56:40 crc kubenswrapper[4818]: I1122 05:56:40.920300 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4j4l" event={"ID":"614051f2-a9cf-4084-9eb0-d5333b5b1d0e","Type":"ContainerDied","Data":"a033544b5969fa584315e0a96787c02ac6d21d90a2eb00e59381f6c76e42cfe6"} Nov 22 05:56:40 crc kubenswrapper[4818]: I1122 05:56:40.920317 4818 scope.go:117] "RemoveContainer" containerID="a4886460724ba5684813c21db17c3989575dea69dff6dcd1076005977210e193" Nov 22 05:56:40 crc kubenswrapper[4818]: I1122 05:56:40.950545 4818 scope.go:117] "RemoveContainer" containerID="2cee89497b77ac47a9bf107b7d4c048076db2e5171713d3612ed4d2831d7ed0c" Nov 22 05:56:40 crc kubenswrapper[4818]: I1122 05:56:40.991741 4818 scope.go:117] "RemoveContainer" containerID="363e7563ae61d8f0492a8522f0937ba006e388d8d0b2cdf2d5c744363c7a55aa" Nov 22 05:56:41 crc kubenswrapper[4818]: I1122 05:56:41.000651 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/614051f2-a9cf-4084-9eb0-d5333b5b1d0e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "614051f2-a9cf-4084-9eb0-d5333b5b1d0e" (UID: "614051f2-a9cf-4084-9eb0-d5333b5b1d0e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:56:41 crc kubenswrapper[4818]: I1122 05:56:41.017167 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/614051f2-a9cf-4084-9eb0-d5333b5b1d0e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 05:56:41 crc kubenswrapper[4818]: I1122 05:56:41.027543 4818 scope.go:117] "RemoveContainer" containerID="a4886460724ba5684813c21db17c3989575dea69dff6dcd1076005977210e193" Nov 22 05:56:41 crc kubenswrapper[4818]: E1122 05:56:41.031615 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4886460724ba5684813c21db17c3989575dea69dff6dcd1076005977210e193\": container with ID starting with a4886460724ba5684813c21db17c3989575dea69dff6dcd1076005977210e193 not found: ID does not exist" containerID="a4886460724ba5684813c21db17c3989575dea69dff6dcd1076005977210e193" Nov 22 05:56:41 crc kubenswrapper[4818]: I1122 05:56:41.031651 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4886460724ba5684813c21db17c3989575dea69dff6dcd1076005977210e193"} err="failed to get container status \"a4886460724ba5684813c21db17c3989575dea69dff6dcd1076005977210e193\": rpc error: code = NotFound desc = could not find container \"a4886460724ba5684813c21db17c3989575dea69dff6dcd1076005977210e193\": container with ID starting with a4886460724ba5684813c21db17c3989575dea69dff6dcd1076005977210e193 not found: ID does not exist" Nov 22 05:56:41 crc kubenswrapper[4818]: I1122 05:56:41.031672 4818 scope.go:117] "RemoveContainer" containerID="2cee89497b77ac47a9bf107b7d4c048076db2e5171713d3612ed4d2831d7ed0c" Nov 22 05:56:41 crc kubenswrapper[4818]: E1122 05:56:41.031962 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cee89497b77ac47a9bf107b7d4c048076db2e5171713d3612ed4d2831d7ed0c\": container with ID starting with 2cee89497b77ac47a9bf107b7d4c048076db2e5171713d3612ed4d2831d7ed0c not found: ID does not exist" containerID="2cee89497b77ac47a9bf107b7d4c048076db2e5171713d3612ed4d2831d7ed0c" Nov 22 05:56:41 crc kubenswrapper[4818]: I1122 05:56:41.032008 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cee89497b77ac47a9bf107b7d4c048076db2e5171713d3612ed4d2831d7ed0c"} err="failed to get container status \"2cee89497b77ac47a9bf107b7d4c048076db2e5171713d3612ed4d2831d7ed0c\": rpc error: code = NotFound desc = could not find container \"2cee89497b77ac47a9bf107b7d4c048076db2e5171713d3612ed4d2831d7ed0c\": container with ID starting with 2cee89497b77ac47a9bf107b7d4c048076db2e5171713d3612ed4d2831d7ed0c not found: ID does not exist" Nov 22 05:56:41 crc kubenswrapper[4818]: I1122 05:56:41.032037 4818 scope.go:117] "RemoveContainer" containerID="363e7563ae61d8f0492a8522f0937ba006e388d8d0b2cdf2d5c744363c7a55aa" Nov 22 05:56:41 crc kubenswrapper[4818]: E1122 05:56:41.032483 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"363e7563ae61d8f0492a8522f0937ba006e388d8d0b2cdf2d5c744363c7a55aa\": container with ID starting with 363e7563ae61d8f0492a8522f0937ba006e388d8d0b2cdf2d5c744363c7a55aa not found: ID does not exist" containerID="363e7563ae61d8f0492a8522f0937ba006e388d8d0b2cdf2d5c744363c7a55aa" Nov 22 05:56:41 crc kubenswrapper[4818]: I1122 05:56:41.032505 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"363e7563ae61d8f0492a8522f0937ba006e388d8d0b2cdf2d5c744363c7a55aa"} err="failed to get container status \"363e7563ae61d8f0492a8522f0937ba006e388d8d0b2cdf2d5c744363c7a55aa\": rpc error: code = NotFound desc = could not find container \"363e7563ae61d8f0492a8522f0937ba006e388d8d0b2cdf2d5c744363c7a55aa\": container with ID starting with 363e7563ae61d8f0492a8522f0937ba006e388d8d0b2cdf2d5c744363c7a55aa not found: ID does not exist" Nov 22 05:56:41 crc kubenswrapper[4818]: I1122 05:56:41.339337 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k4j4l"] Nov 22 05:56:41 crc kubenswrapper[4818]: I1122 05:56:41.351152 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k4j4l"] Nov 22 05:56:42 crc kubenswrapper[4818]: I1122 05:56:42.305533 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="614051f2-a9cf-4084-9eb0-d5333b5b1d0e" path="/var/lib/kubelet/pods/614051f2-a9cf-4084-9eb0-d5333b5b1d0e/volumes" Nov 22 05:56:43 crc kubenswrapper[4818]: I1122 05:56:43.454567 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lwcql" Nov 22 05:56:44 crc kubenswrapper[4818]: I1122 05:56:44.268641 4818 scope.go:117] "RemoveContainer" containerID="efd6a02edfb01d2064fc0e366c29accec0a100d9d2b56e5807f6c968832182b8" Nov 22 05:56:45 crc kubenswrapper[4818]: I1122 05:56:45.705878 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lwcql"] Nov 22 05:56:45 crc kubenswrapper[4818]: I1122 05:56:45.714849 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lwcql" podUID="72a97bea-43fc-40b7-abd0-854197951ac8" containerName="registry-server" containerID="cri-o://f3a4d787a2bf0a4ec2a1594251294a8a5472566da9704fba73961798116a4921" gracePeriod=2 Nov 22 05:56:45 crc kubenswrapper[4818]: I1122 05:56:45.998277 4818 generic.go:334] "Generic (PLEG): container finished" podID="72a97bea-43fc-40b7-abd0-854197951ac8" containerID="f3a4d787a2bf0a4ec2a1594251294a8a5472566da9704fba73961798116a4921" exitCode=0 Nov 22 05:56:45 crc kubenswrapper[4818]: I1122 05:56:45.998441 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lwcql" event={"ID":"72a97bea-43fc-40b7-abd0-854197951ac8","Type":"ContainerDied","Data":"f3a4d787a2bf0a4ec2a1594251294a8a5472566da9704fba73961798116a4921"} Nov 22 05:56:46 crc kubenswrapper[4818]: I1122 05:56:46.314606 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lwcql" Nov 22 05:56:46 crc kubenswrapper[4818]: I1122 05:56:46.465114 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwxwm\" (UniqueName: \"kubernetes.io/projected/72a97bea-43fc-40b7-abd0-854197951ac8-kube-api-access-wwxwm\") pod \"72a97bea-43fc-40b7-abd0-854197951ac8\" (UID: \"72a97bea-43fc-40b7-abd0-854197951ac8\") " Nov 22 05:56:46 crc kubenswrapper[4818]: I1122 05:56:46.465244 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72a97bea-43fc-40b7-abd0-854197951ac8-catalog-content\") pod \"72a97bea-43fc-40b7-abd0-854197951ac8\" (UID: \"72a97bea-43fc-40b7-abd0-854197951ac8\") " Nov 22 05:56:46 crc kubenswrapper[4818]: I1122 05:56:46.465482 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72a97bea-43fc-40b7-abd0-854197951ac8-utilities\") pod \"72a97bea-43fc-40b7-abd0-854197951ac8\" (UID: \"72a97bea-43fc-40b7-abd0-854197951ac8\") " Nov 22 05:56:46 crc kubenswrapper[4818]: I1122 05:56:46.466469 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72a97bea-43fc-40b7-abd0-854197951ac8-utilities" (OuterVolumeSpecName: "utilities") pod "72a97bea-43fc-40b7-abd0-854197951ac8" (UID: "72a97bea-43fc-40b7-abd0-854197951ac8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:56:46 crc kubenswrapper[4818]: I1122 05:56:46.483768 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72a97bea-43fc-40b7-abd0-854197951ac8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72a97bea-43fc-40b7-abd0-854197951ac8" (UID: "72a97bea-43fc-40b7-abd0-854197951ac8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 05:56:46 crc kubenswrapper[4818]: I1122 05:56:46.567577 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72a97bea-43fc-40b7-abd0-854197951ac8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 05:56:46 crc kubenswrapper[4818]: I1122 05:56:46.567616 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72a97bea-43fc-40b7-abd0-854197951ac8-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 05:56:47 crc kubenswrapper[4818]: I1122 05:56:47.010950 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lwcql" event={"ID":"72a97bea-43fc-40b7-abd0-854197951ac8","Type":"ContainerDied","Data":"2ba3d5e7628203146a6a6dc0d7248963636093a339195a52bab879140b4725c0"} Nov 22 05:56:47 crc kubenswrapper[4818]: I1122 05:56:47.011011 4818 scope.go:117] "RemoveContainer" containerID="f3a4d787a2bf0a4ec2a1594251294a8a5472566da9704fba73961798116a4921" Nov 22 05:56:47 crc kubenswrapper[4818]: I1122 05:56:47.011032 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lwcql" Nov 22 05:56:47 crc kubenswrapper[4818]: I1122 05:56:47.046740 4818 scope.go:117] "RemoveContainer" containerID="6ea72e9bfee027b7f049dd5edb02d3f848f70aecda4c13938a28e3a50efaf00b" Nov 22 05:56:47 crc kubenswrapper[4818]: I1122 05:56:47.093982 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72a97bea-43fc-40b7-abd0-854197951ac8-kube-api-access-wwxwm" (OuterVolumeSpecName: "kube-api-access-wwxwm") pod "72a97bea-43fc-40b7-abd0-854197951ac8" (UID: "72a97bea-43fc-40b7-abd0-854197951ac8"). InnerVolumeSpecName "kube-api-access-wwxwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 05:56:47 crc kubenswrapper[4818]: I1122 05:56:47.110167 4818 scope.go:117] "RemoveContainer" containerID="c7273c8003724742709026453738c67574159c11278e521bdc50e80bc06d819d" Nov 22 05:56:47 crc kubenswrapper[4818]: I1122 05:56:47.180179 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwxwm\" (UniqueName: \"kubernetes.io/projected/72a97bea-43fc-40b7-abd0-854197951ac8-kube-api-access-wwxwm\") on node \"crc\" DevicePath \"\"" Nov 22 05:56:47 crc kubenswrapper[4818]: I1122 05:56:47.347762 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lwcql"] Nov 22 05:56:47 crc kubenswrapper[4818]: I1122 05:56:47.355616 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lwcql"] Nov 22 05:56:48 crc kubenswrapper[4818]: I1122 05:56:48.313524 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72a97bea-43fc-40b7-abd0-854197951ac8" path="/var/lib/kubelet/pods/72a97bea-43fc-40b7-abd0-854197951ac8/volumes" Nov 22 05:57:21 crc kubenswrapper[4818]: I1122 05:57:21.264966 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:57:21 crc kubenswrapper[4818]: I1122 05:57:21.265746 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:57:51 crc kubenswrapper[4818]: I1122 05:57:51.264749 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:57:51 crc kubenswrapper[4818]: I1122 05:57:51.265498 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:58:21 crc kubenswrapper[4818]: I1122 05:58:21.264623 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 05:58:21 crc kubenswrapper[4818]: I1122 05:58:21.265442 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 05:58:21 crc kubenswrapper[4818]: I1122 05:58:21.265543 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 05:58:21 crc kubenswrapper[4818]: I1122 05:58:21.266948 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"87e53488d73d6090d237f383bc00e58356c78446d810b06ded7d6522cfcda586"} pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 05:58:21 crc kubenswrapper[4818]: I1122 05:58:21.267083 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" containerID="cri-o://87e53488d73d6090d237f383bc00e58356c78446d810b06ded7d6522cfcda586" gracePeriod=600 Nov 22 05:58:22 crc kubenswrapper[4818]: I1122 05:58:22.075165 4818 generic.go:334] "Generic (PLEG): container finished" podID="0971c9ec-29ce-44eb-9875-ad4664143305" containerID="87e53488d73d6090d237f383bc00e58356c78446d810b06ded7d6522cfcda586" exitCode=0 Nov 22 05:58:22 crc kubenswrapper[4818]: I1122 05:58:22.075238 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerDied","Data":"87e53488d73d6090d237f383bc00e58356c78446d810b06ded7d6522cfcda586"} Nov 22 05:58:22 crc kubenswrapper[4818]: I1122 05:58:22.076011 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861"} Nov 22 05:58:22 crc kubenswrapper[4818]: I1122 05:58:22.076072 4818 scope.go:117] "RemoveContainer" containerID="5a7dac167f0dea07bf6c373a0cefe8811a0e92fc0be637f26e7103349bc57849" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.145484 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396520-pfm2q"] Nov 22 06:00:00 crc kubenswrapper[4818]: E1122 06:00:00.146390 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72a97bea-43fc-40b7-abd0-854197951ac8" containerName="extract-content" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.146403 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="72a97bea-43fc-40b7-abd0-854197951ac8" containerName="extract-content" Nov 22 06:00:00 crc kubenswrapper[4818]: E1122 06:00:00.146440 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="614051f2-a9cf-4084-9eb0-d5333b5b1d0e" containerName="extract-utilities" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.146447 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="614051f2-a9cf-4084-9eb0-d5333b5b1d0e" containerName="extract-utilities" Nov 22 06:00:00 crc kubenswrapper[4818]: E1122 06:00:00.146454 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72a97bea-43fc-40b7-abd0-854197951ac8" containerName="extract-utilities" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.146461 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="72a97bea-43fc-40b7-abd0-854197951ac8" containerName="extract-utilities" Nov 22 06:00:00 crc kubenswrapper[4818]: E1122 06:00:00.146474 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="614051f2-a9cf-4084-9eb0-d5333b5b1d0e" containerName="registry-server" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.146480 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="614051f2-a9cf-4084-9eb0-d5333b5b1d0e" containerName="registry-server" Nov 22 06:00:00 crc kubenswrapper[4818]: E1122 06:00:00.146498 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72a97bea-43fc-40b7-abd0-854197951ac8" containerName="registry-server" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.146503 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="72a97bea-43fc-40b7-abd0-854197951ac8" containerName="registry-server" Nov 22 06:00:00 crc kubenswrapper[4818]: E1122 06:00:00.146514 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="614051f2-a9cf-4084-9eb0-d5333b5b1d0e" containerName="extract-content" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.146520 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="614051f2-a9cf-4084-9eb0-d5333b5b1d0e" containerName="extract-content" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.146678 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="72a97bea-43fc-40b7-abd0-854197951ac8" containerName="registry-server" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.146702 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="614051f2-a9cf-4084-9eb0-d5333b5b1d0e" containerName="registry-server" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.147349 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396520-pfm2q" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.149279 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.149695 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.164240 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396520-pfm2q"] Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.202038 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g97gq\" (UniqueName: \"kubernetes.io/projected/0b54cb44-82b4-4565-82bc-d0fc466e3f0b-kube-api-access-g97gq\") pod \"collect-profiles-29396520-pfm2q\" (UID: \"0b54cb44-82b4-4565-82bc-d0fc466e3f0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396520-pfm2q" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.202144 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b54cb44-82b4-4565-82bc-d0fc466e3f0b-config-volume\") pod \"collect-profiles-29396520-pfm2q\" (UID: \"0b54cb44-82b4-4565-82bc-d0fc466e3f0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396520-pfm2q" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.202223 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b54cb44-82b4-4565-82bc-d0fc466e3f0b-secret-volume\") pod \"collect-profiles-29396520-pfm2q\" (UID: \"0b54cb44-82b4-4565-82bc-d0fc466e3f0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396520-pfm2q" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.304160 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b54cb44-82b4-4565-82bc-d0fc466e3f0b-config-volume\") pod \"collect-profiles-29396520-pfm2q\" (UID: \"0b54cb44-82b4-4565-82bc-d0fc466e3f0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396520-pfm2q" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.304292 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b54cb44-82b4-4565-82bc-d0fc466e3f0b-secret-volume\") pod \"collect-profiles-29396520-pfm2q\" (UID: \"0b54cb44-82b4-4565-82bc-d0fc466e3f0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396520-pfm2q" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.304373 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g97gq\" (UniqueName: \"kubernetes.io/projected/0b54cb44-82b4-4565-82bc-d0fc466e3f0b-kube-api-access-g97gq\") pod \"collect-profiles-29396520-pfm2q\" (UID: \"0b54cb44-82b4-4565-82bc-d0fc466e3f0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396520-pfm2q" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.305291 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b54cb44-82b4-4565-82bc-d0fc466e3f0b-config-volume\") pod \"collect-profiles-29396520-pfm2q\" (UID: \"0b54cb44-82b4-4565-82bc-d0fc466e3f0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396520-pfm2q" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.310505 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b54cb44-82b4-4565-82bc-d0fc466e3f0b-secret-volume\") pod \"collect-profiles-29396520-pfm2q\" (UID: \"0b54cb44-82b4-4565-82bc-d0fc466e3f0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396520-pfm2q" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.318443 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g97gq\" (UniqueName: \"kubernetes.io/projected/0b54cb44-82b4-4565-82bc-d0fc466e3f0b-kube-api-access-g97gq\") pod \"collect-profiles-29396520-pfm2q\" (UID: \"0b54cb44-82b4-4565-82bc-d0fc466e3f0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396520-pfm2q" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.471886 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396520-pfm2q" Nov 22 06:00:00 crc kubenswrapper[4818]: I1122 06:00:00.960345 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396520-pfm2q"] Nov 22 06:00:01 crc kubenswrapper[4818]: I1122 06:00:01.126955 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396520-pfm2q" event={"ID":"0b54cb44-82b4-4565-82bc-d0fc466e3f0b","Type":"ContainerStarted","Data":"192ccd4a0321706af6b52cf21b63b61ebc52ea188450c6ce9e6af293dce6fbe0"} Nov 22 06:00:02 crc kubenswrapper[4818]: I1122 06:00:02.134846 4818 generic.go:334] "Generic (PLEG): container finished" podID="0b54cb44-82b4-4565-82bc-d0fc466e3f0b" containerID="2a81588539c4b704a021ecc251155922963845e86a954ee753383643681a1095" exitCode=0 Nov 22 06:00:02 crc kubenswrapper[4818]: I1122 06:00:02.134905 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396520-pfm2q" event={"ID":"0b54cb44-82b4-4565-82bc-d0fc466e3f0b","Type":"ContainerDied","Data":"2a81588539c4b704a021ecc251155922963845e86a954ee753383643681a1095"} Nov 22 06:00:03 crc kubenswrapper[4818]: I1122 06:00:03.533689 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396520-pfm2q" Nov 22 06:00:03 crc kubenswrapper[4818]: I1122 06:00:03.575374 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b54cb44-82b4-4565-82bc-d0fc466e3f0b-config-volume\") pod \"0b54cb44-82b4-4565-82bc-d0fc466e3f0b\" (UID: \"0b54cb44-82b4-4565-82bc-d0fc466e3f0b\") " Nov 22 06:00:03 crc kubenswrapper[4818]: I1122 06:00:03.575434 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g97gq\" (UniqueName: \"kubernetes.io/projected/0b54cb44-82b4-4565-82bc-d0fc466e3f0b-kube-api-access-g97gq\") pod \"0b54cb44-82b4-4565-82bc-d0fc466e3f0b\" (UID: \"0b54cb44-82b4-4565-82bc-d0fc466e3f0b\") " Nov 22 06:00:03 crc kubenswrapper[4818]: I1122 06:00:03.575519 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b54cb44-82b4-4565-82bc-d0fc466e3f0b-secret-volume\") pod \"0b54cb44-82b4-4565-82bc-d0fc466e3f0b\" (UID: \"0b54cb44-82b4-4565-82bc-d0fc466e3f0b\") " Nov 22 06:00:03 crc kubenswrapper[4818]: I1122 06:00:03.576220 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b54cb44-82b4-4565-82bc-d0fc466e3f0b-config-volume" (OuterVolumeSpecName: "config-volume") pod "0b54cb44-82b4-4565-82bc-d0fc466e3f0b" (UID: "0b54cb44-82b4-4565-82bc-d0fc466e3f0b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 06:00:03 crc kubenswrapper[4818]: I1122 06:00:03.581131 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b54cb44-82b4-4565-82bc-d0fc466e3f0b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0b54cb44-82b4-4565-82bc-d0fc466e3f0b" (UID: "0b54cb44-82b4-4565-82bc-d0fc466e3f0b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 06:00:03 crc kubenswrapper[4818]: I1122 06:00:03.581609 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b54cb44-82b4-4565-82bc-d0fc466e3f0b-kube-api-access-g97gq" (OuterVolumeSpecName: "kube-api-access-g97gq") pod "0b54cb44-82b4-4565-82bc-d0fc466e3f0b" (UID: "0b54cb44-82b4-4565-82bc-d0fc466e3f0b"). InnerVolumeSpecName "kube-api-access-g97gq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 06:00:03 crc kubenswrapper[4818]: I1122 06:00:03.678368 4818 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b54cb44-82b4-4565-82bc-d0fc466e3f0b-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 06:00:03 crc kubenswrapper[4818]: I1122 06:00:03.678408 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g97gq\" (UniqueName: \"kubernetes.io/projected/0b54cb44-82b4-4565-82bc-d0fc466e3f0b-kube-api-access-g97gq\") on node \"crc\" DevicePath \"\"" Nov 22 06:00:03 crc kubenswrapper[4818]: I1122 06:00:03.678421 4818 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b54cb44-82b4-4565-82bc-d0fc466e3f0b-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 06:00:04 crc kubenswrapper[4818]: I1122 06:00:04.152152 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396520-pfm2q" event={"ID":"0b54cb44-82b4-4565-82bc-d0fc466e3f0b","Type":"ContainerDied","Data":"192ccd4a0321706af6b52cf21b63b61ebc52ea188450c6ce9e6af293dce6fbe0"} Nov 22 06:00:04 crc kubenswrapper[4818]: I1122 06:00:04.152490 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="192ccd4a0321706af6b52cf21b63b61ebc52ea188450c6ce9e6af293dce6fbe0" Nov 22 06:00:04 crc kubenswrapper[4818]: I1122 06:00:04.152204 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396520-pfm2q" Nov 22 06:00:04 crc kubenswrapper[4818]: I1122 06:00:04.630346 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w"] Nov 22 06:00:04 crc kubenswrapper[4818]: I1122 06:00:04.635525 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396475-zcs8w"] Nov 22 06:00:06 crc kubenswrapper[4818]: I1122 06:00:06.304339 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33d93220-7194-4413-9d83-3eb08fe261d4" path="/var/lib/kubelet/pods/33d93220-7194-4413-9d83-3eb08fe261d4/volumes" Nov 22 06:00:21 crc kubenswrapper[4818]: I1122 06:00:21.264801 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 06:00:21 crc kubenswrapper[4818]: I1122 06:00:21.265471 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 06:00:24 crc kubenswrapper[4818]: I1122 06:00:24.363694 4818 generic.go:334] "Generic (PLEG): container finished" podID="df264e71-c1de-4e1f-9a56-24f5bd284a80" containerID="938560beedaa5d2f146e233c4f8cdd631f93c381aba0bfe9cf75146edf77c60f" exitCode=0 Nov 22 06:00:24 crc kubenswrapper[4818]: I1122 06:00:24.363835 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"df264e71-c1de-4e1f-9a56-24f5bd284a80","Type":"ContainerDied","Data":"938560beedaa5d2f146e233c4f8cdd631f93c381aba0bfe9cf75146edf77c60f"} Nov 22 06:00:25 crc kubenswrapper[4818]: I1122 06:00:25.882002 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 22 06:00:25 crc kubenswrapper[4818]: I1122 06:00:25.989417 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/df264e71-c1de-4e1f-9a56-24f5bd284a80-test-operator-ephemeral-temporary\") pod \"df264e71-c1de-4e1f-9a56-24f5bd284a80\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " Nov 22 06:00:25 crc kubenswrapper[4818]: I1122 06:00:25.989520 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cw2nn\" (UniqueName: \"kubernetes.io/projected/df264e71-c1de-4e1f-9a56-24f5bd284a80-kube-api-access-cw2nn\") pod \"df264e71-c1de-4e1f-9a56-24f5bd284a80\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " Nov 22 06:00:25 crc kubenswrapper[4818]: I1122 06:00:25.989679 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"df264e71-c1de-4e1f-9a56-24f5bd284a80\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " Nov 22 06:00:25 crc kubenswrapper[4818]: I1122 06:00:25.989720 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df264e71-c1de-4e1f-9a56-24f5bd284a80-config-data\") pod \"df264e71-c1de-4e1f-9a56-24f5bd284a80\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " Nov 22 06:00:25 crc kubenswrapper[4818]: I1122 06:00:25.989748 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/df264e71-c1de-4e1f-9a56-24f5bd284a80-ca-certs\") pod \"df264e71-c1de-4e1f-9a56-24f5bd284a80\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " Nov 22 06:00:25 crc kubenswrapper[4818]: I1122 06:00:25.989799 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/df264e71-c1de-4e1f-9a56-24f5bd284a80-openstack-config\") pod \"df264e71-c1de-4e1f-9a56-24f5bd284a80\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " Nov 22 06:00:25 crc kubenswrapper[4818]: I1122 06:00:25.989843 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df264e71-c1de-4e1f-9a56-24f5bd284a80-ssh-key\") pod \"df264e71-c1de-4e1f-9a56-24f5bd284a80\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " Nov 22 06:00:25 crc kubenswrapper[4818]: I1122 06:00:25.990131 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df264e71-c1de-4e1f-9a56-24f5bd284a80-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "df264e71-c1de-4e1f-9a56-24f5bd284a80" (UID: "df264e71-c1de-4e1f-9a56-24f5bd284a80"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 06:00:25 crc kubenswrapper[4818]: I1122 06:00:25.990722 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df264e71-c1de-4e1f-9a56-24f5bd284a80-config-data" (OuterVolumeSpecName: "config-data") pod "df264e71-c1de-4e1f-9a56-24f5bd284a80" (UID: "df264e71-c1de-4e1f-9a56-24f5bd284a80"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 06:00:25 crc kubenswrapper[4818]: I1122 06:00:25.991281 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/df264e71-c1de-4e1f-9a56-24f5bd284a80-test-operator-ephemeral-workdir\") pod \"df264e71-c1de-4e1f-9a56-24f5bd284a80\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " Nov 22 06:00:25 crc kubenswrapper[4818]: I1122 06:00:25.991332 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/df264e71-c1de-4e1f-9a56-24f5bd284a80-openstack-config-secret\") pod \"df264e71-c1de-4e1f-9a56-24f5bd284a80\" (UID: \"df264e71-c1de-4e1f-9a56-24f5bd284a80\") " Nov 22 06:00:25 crc kubenswrapper[4818]: I1122 06:00:25.992122 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df264e71-c1de-4e1f-9a56-24f5bd284a80-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 06:00:25 crc kubenswrapper[4818]: I1122 06:00:25.992149 4818 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/df264e71-c1de-4e1f-9a56-24f5bd284a80-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 22 06:00:25 crc kubenswrapper[4818]: I1122 06:00:25.996227 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df264e71-c1de-4e1f-9a56-24f5bd284a80-kube-api-access-cw2nn" (OuterVolumeSpecName: "kube-api-access-cw2nn") pod "df264e71-c1de-4e1f-9a56-24f5bd284a80" (UID: "df264e71-c1de-4e1f-9a56-24f5bd284a80"). InnerVolumeSpecName "kube-api-access-cw2nn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 06:00:25 crc kubenswrapper[4818]: I1122 06:00:25.999438 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df264e71-c1de-4e1f-9a56-24f5bd284a80-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "df264e71-c1de-4e1f-9a56-24f5bd284a80" (UID: "df264e71-c1de-4e1f-9a56-24f5bd284a80"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 06:00:25 crc kubenswrapper[4818]: I1122 06:00:25.999592 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "test-operator-logs") pod "df264e71-c1de-4e1f-9a56-24f5bd284a80" (UID: "df264e71-c1de-4e1f-9a56-24f5bd284a80"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 06:00:26 crc kubenswrapper[4818]: I1122 06:00:26.025320 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df264e71-c1de-4e1f-9a56-24f5bd284a80-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "df264e71-c1de-4e1f-9a56-24f5bd284a80" (UID: "df264e71-c1de-4e1f-9a56-24f5bd284a80"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 06:00:26 crc kubenswrapper[4818]: I1122 06:00:26.041718 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df264e71-c1de-4e1f-9a56-24f5bd284a80-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "df264e71-c1de-4e1f-9a56-24f5bd284a80" (UID: "df264e71-c1de-4e1f-9a56-24f5bd284a80"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 06:00:26 crc kubenswrapper[4818]: I1122 06:00:26.044406 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df264e71-c1de-4e1f-9a56-24f5bd284a80-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "df264e71-c1de-4e1f-9a56-24f5bd284a80" (UID: "df264e71-c1de-4e1f-9a56-24f5bd284a80"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 06:00:26 crc kubenswrapper[4818]: I1122 06:00:26.048902 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df264e71-c1de-4e1f-9a56-24f5bd284a80-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "df264e71-c1de-4e1f-9a56-24f5bd284a80" (UID: "df264e71-c1de-4e1f-9a56-24f5bd284a80"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 06:00:26 crc kubenswrapper[4818]: I1122 06:00:26.094464 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cw2nn\" (UniqueName: \"kubernetes.io/projected/df264e71-c1de-4e1f-9a56-24f5bd284a80-kube-api-access-cw2nn\") on node \"crc\" DevicePath \"\"" Nov 22 06:00:26 crc kubenswrapper[4818]: I1122 06:00:26.094531 4818 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 22 06:00:26 crc kubenswrapper[4818]: I1122 06:00:26.094546 4818 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/df264e71-c1de-4e1f-9a56-24f5bd284a80-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 22 06:00:26 crc kubenswrapper[4818]: I1122 06:00:26.094560 4818 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/df264e71-c1de-4e1f-9a56-24f5bd284a80-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 22 06:00:26 crc kubenswrapper[4818]: I1122 06:00:26.094572 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df264e71-c1de-4e1f-9a56-24f5bd284a80-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 06:00:26 crc kubenswrapper[4818]: I1122 06:00:26.094584 4818 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/df264e71-c1de-4e1f-9a56-24f5bd284a80-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 22 06:00:26 crc kubenswrapper[4818]: I1122 06:00:26.094597 4818 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/df264e71-c1de-4e1f-9a56-24f5bd284a80-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 22 06:00:26 crc kubenswrapper[4818]: I1122 06:00:26.122854 4818 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 22 06:00:26 crc kubenswrapper[4818]: I1122 06:00:26.196923 4818 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 22 06:00:26 crc kubenswrapper[4818]: I1122 06:00:26.387052 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"df264e71-c1de-4e1f-9a56-24f5bd284a80","Type":"ContainerDied","Data":"ef0e41d6ee637b3a795992b3238c7cdec4790c79ff7269354ce57917c414deab"} Nov 22 06:00:26 crc kubenswrapper[4818]: I1122 06:00:26.387128 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef0e41d6ee637b3a795992b3238c7cdec4790c79ff7269354ce57917c414deab" Nov 22 06:00:26 crc kubenswrapper[4818]: I1122 06:00:26.387136 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 22 06:00:38 crc kubenswrapper[4818]: I1122 06:00:38.687352 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 22 06:00:38 crc kubenswrapper[4818]: E1122 06:00:38.688514 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b54cb44-82b4-4565-82bc-d0fc466e3f0b" containerName="collect-profiles" Nov 22 06:00:38 crc kubenswrapper[4818]: I1122 06:00:38.688530 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b54cb44-82b4-4565-82bc-d0fc466e3f0b" containerName="collect-profiles" Nov 22 06:00:38 crc kubenswrapper[4818]: E1122 06:00:38.688547 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df264e71-c1de-4e1f-9a56-24f5bd284a80" containerName="tempest-tests-tempest-tests-runner" Nov 22 06:00:38 crc kubenswrapper[4818]: I1122 06:00:38.688554 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="df264e71-c1de-4e1f-9a56-24f5bd284a80" containerName="tempest-tests-tempest-tests-runner" Nov 22 06:00:38 crc kubenswrapper[4818]: I1122 06:00:38.688804 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="df264e71-c1de-4e1f-9a56-24f5bd284a80" containerName="tempest-tests-tempest-tests-runner" Nov 22 06:00:38 crc kubenswrapper[4818]: I1122 06:00:38.688822 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b54cb44-82b4-4565-82bc-d0fc466e3f0b" containerName="collect-profiles" Nov 22 06:00:38 crc kubenswrapper[4818]: I1122 06:00:38.689685 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 06:00:38 crc kubenswrapper[4818]: I1122 06:00:38.693861 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-9vz8j" Nov 22 06:00:38 crc kubenswrapper[4818]: I1122 06:00:38.704765 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 22 06:00:38 crc kubenswrapper[4818]: I1122 06:00:38.792984 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5xv7\" (UniqueName: \"kubernetes.io/projected/fce70419-9c81-461c-bcb5-e829c1868e03-kube-api-access-b5xv7\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fce70419-9c81-461c-bcb5-e829c1868e03\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 06:00:38 crc kubenswrapper[4818]: I1122 06:00:38.793202 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fce70419-9c81-461c-bcb5-e829c1868e03\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 06:00:38 crc kubenswrapper[4818]: I1122 06:00:38.895538 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5xv7\" (UniqueName: \"kubernetes.io/projected/fce70419-9c81-461c-bcb5-e829c1868e03-kube-api-access-b5xv7\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fce70419-9c81-461c-bcb5-e829c1868e03\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 06:00:38 crc kubenswrapper[4818]: I1122 06:00:38.895712 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fce70419-9c81-461c-bcb5-e829c1868e03\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 06:00:38 crc kubenswrapper[4818]: I1122 06:00:38.898006 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fce70419-9c81-461c-bcb5-e829c1868e03\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 06:00:38 crc kubenswrapper[4818]: I1122 06:00:38.925138 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5xv7\" (UniqueName: \"kubernetes.io/projected/fce70419-9c81-461c-bcb5-e829c1868e03-kube-api-access-b5xv7\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fce70419-9c81-461c-bcb5-e829c1868e03\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 06:00:38 crc kubenswrapper[4818]: I1122 06:00:38.933996 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fce70419-9c81-461c-bcb5-e829c1868e03\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 06:00:39 crc kubenswrapper[4818]: I1122 06:00:39.021763 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 06:00:39 crc kubenswrapper[4818]: I1122 06:00:39.521677 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 22 06:00:39 crc kubenswrapper[4818]: I1122 06:00:39.529058 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 06:00:40 crc kubenswrapper[4818]: I1122 06:00:40.542200 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"fce70419-9c81-461c-bcb5-e829c1868e03","Type":"ContainerStarted","Data":"c84b32adf311a03fe025e79721350f325ecd6171a8bc29865a459220841b047f"} Nov 22 06:00:41 crc kubenswrapper[4818]: I1122 06:00:41.565959 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"fce70419-9c81-461c-bcb5-e829c1868e03","Type":"ContainerStarted","Data":"157f46549952d43570d355169273e18ced31a80712ba44683a421bf124398335"} Nov 22 06:00:41 crc kubenswrapper[4818]: I1122 06:00:41.584647 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.665765528 podStartE2EDuration="3.58459496s" podCreationTimestamp="2025-11-22 06:00:38 +0000 UTC" firstStartedPulling="2025-11-22 06:00:39.528721938 +0000 UTC m=+4392.103138485" lastFinishedPulling="2025-11-22 06:00:40.44755138 +0000 UTC m=+4393.021967917" observedRunningTime="2025-11-22 06:00:41.582989537 +0000 UTC m=+4394.157406074" watchObservedRunningTime="2025-11-22 06:00:41.58459496 +0000 UTC m=+4394.159011527" Nov 22 06:00:44 crc kubenswrapper[4818]: I1122 06:00:44.508126 4818 scope.go:117] "RemoveContainer" containerID="33b0c0a8d76db88f05bf6e94e1d03cce8a2ce5ec1aaf25dd070ac8633ed03fd2" Nov 22 06:00:51 crc kubenswrapper[4818]: I1122 06:00:51.264777 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 06:00:51 crc kubenswrapper[4818]: I1122 06:00:51.265403 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 06:01:00 crc kubenswrapper[4818]: I1122 06:01:00.166717 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29396521-h4m77"] Nov 22 06:01:00 crc kubenswrapper[4818]: I1122 06:01:00.168859 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29396521-h4m77" Nov 22 06:01:00 crc kubenswrapper[4818]: I1122 06:01:00.178606 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29396521-h4m77"] Nov 22 06:01:00 crc kubenswrapper[4818]: I1122 06:01:00.255300 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/be3d9b50-d430-4a39-85e9-3afb138cfdf4-fernet-keys\") pod \"keystone-cron-29396521-h4m77\" (UID: \"be3d9b50-d430-4a39-85e9-3afb138cfdf4\") " pod="openstack/keystone-cron-29396521-h4m77" Nov 22 06:01:00 crc kubenswrapper[4818]: I1122 06:01:00.255375 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfn76\" (UniqueName: \"kubernetes.io/projected/be3d9b50-d430-4a39-85e9-3afb138cfdf4-kube-api-access-jfn76\") pod \"keystone-cron-29396521-h4m77\" (UID: \"be3d9b50-d430-4a39-85e9-3afb138cfdf4\") " pod="openstack/keystone-cron-29396521-h4m77" Nov 22 06:01:00 crc kubenswrapper[4818]: I1122 06:01:00.255400 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be3d9b50-d430-4a39-85e9-3afb138cfdf4-config-data\") pod \"keystone-cron-29396521-h4m77\" (UID: \"be3d9b50-d430-4a39-85e9-3afb138cfdf4\") " pod="openstack/keystone-cron-29396521-h4m77" Nov 22 06:01:00 crc kubenswrapper[4818]: I1122 06:01:00.255748 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be3d9b50-d430-4a39-85e9-3afb138cfdf4-combined-ca-bundle\") pod \"keystone-cron-29396521-h4m77\" (UID: \"be3d9b50-d430-4a39-85e9-3afb138cfdf4\") " pod="openstack/keystone-cron-29396521-h4m77" Nov 22 06:01:00 crc kubenswrapper[4818]: I1122 06:01:00.357071 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be3d9b50-d430-4a39-85e9-3afb138cfdf4-combined-ca-bundle\") pod \"keystone-cron-29396521-h4m77\" (UID: \"be3d9b50-d430-4a39-85e9-3afb138cfdf4\") " pod="openstack/keystone-cron-29396521-h4m77" Nov 22 06:01:00 crc kubenswrapper[4818]: I1122 06:01:00.357212 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/be3d9b50-d430-4a39-85e9-3afb138cfdf4-fernet-keys\") pod \"keystone-cron-29396521-h4m77\" (UID: \"be3d9b50-d430-4a39-85e9-3afb138cfdf4\") " pod="openstack/keystone-cron-29396521-h4m77" Nov 22 06:01:00 crc kubenswrapper[4818]: I1122 06:01:00.357293 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfn76\" (UniqueName: \"kubernetes.io/projected/be3d9b50-d430-4a39-85e9-3afb138cfdf4-kube-api-access-jfn76\") pod \"keystone-cron-29396521-h4m77\" (UID: \"be3d9b50-d430-4a39-85e9-3afb138cfdf4\") " pod="openstack/keystone-cron-29396521-h4m77" Nov 22 06:01:00 crc kubenswrapper[4818]: I1122 06:01:00.357323 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be3d9b50-d430-4a39-85e9-3afb138cfdf4-config-data\") pod \"keystone-cron-29396521-h4m77\" (UID: \"be3d9b50-d430-4a39-85e9-3afb138cfdf4\") " pod="openstack/keystone-cron-29396521-h4m77" Nov 22 06:01:00 crc kubenswrapper[4818]: I1122 06:01:00.596542 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/be3d9b50-d430-4a39-85e9-3afb138cfdf4-fernet-keys\") pod \"keystone-cron-29396521-h4m77\" (UID: \"be3d9b50-d430-4a39-85e9-3afb138cfdf4\") " pod="openstack/keystone-cron-29396521-h4m77" Nov 22 06:01:00 crc kubenswrapper[4818]: I1122 06:01:00.597244 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfn76\" (UniqueName: \"kubernetes.io/projected/be3d9b50-d430-4a39-85e9-3afb138cfdf4-kube-api-access-jfn76\") pod \"keystone-cron-29396521-h4m77\" (UID: \"be3d9b50-d430-4a39-85e9-3afb138cfdf4\") " pod="openstack/keystone-cron-29396521-h4m77" Nov 22 06:01:00 crc kubenswrapper[4818]: I1122 06:01:00.597487 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be3d9b50-d430-4a39-85e9-3afb138cfdf4-combined-ca-bundle\") pod \"keystone-cron-29396521-h4m77\" (UID: \"be3d9b50-d430-4a39-85e9-3afb138cfdf4\") " pod="openstack/keystone-cron-29396521-h4m77" Nov 22 06:01:00 crc kubenswrapper[4818]: I1122 06:01:00.598529 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be3d9b50-d430-4a39-85e9-3afb138cfdf4-config-data\") pod \"keystone-cron-29396521-h4m77\" (UID: \"be3d9b50-d430-4a39-85e9-3afb138cfdf4\") " pod="openstack/keystone-cron-29396521-h4m77" Nov 22 06:01:00 crc kubenswrapper[4818]: I1122 06:01:00.798985 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29396521-h4m77" Nov 22 06:01:01 crc kubenswrapper[4818]: I1122 06:01:01.267868 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29396521-h4m77"] Nov 22 06:01:01 crc kubenswrapper[4818]: I1122 06:01:01.772969 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29396521-h4m77" event={"ID":"be3d9b50-d430-4a39-85e9-3afb138cfdf4","Type":"ContainerStarted","Data":"a45ed19a0c9a6f1cab029647f675d9156f5723ea63dd07a4ac4aecd8e9992509"} Nov 22 06:01:01 crc kubenswrapper[4818]: I1122 06:01:01.773364 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29396521-h4m77" event={"ID":"be3d9b50-d430-4a39-85e9-3afb138cfdf4","Type":"ContainerStarted","Data":"6fb2750d99b14c54a8d424d922c4419f545d415b23372eb5625924c95a0f049d"} Nov 22 06:01:01 crc kubenswrapper[4818]: I1122 06:01:01.801382 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29396521-h4m77" podStartSLOduration=1.8013651899999998 podStartE2EDuration="1.80136519s" podCreationTimestamp="2025-11-22 06:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 06:01:01.794426165 +0000 UTC m=+4414.368842712" watchObservedRunningTime="2025-11-22 06:01:01.80136519 +0000 UTC m=+4414.375781717" Nov 22 06:01:05 crc kubenswrapper[4818]: I1122 06:01:05.845961 4818 generic.go:334] "Generic (PLEG): container finished" podID="be3d9b50-d430-4a39-85e9-3afb138cfdf4" containerID="a45ed19a0c9a6f1cab029647f675d9156f5723ea63dd07a4ac4aecd8e9992509" exitCode=0 Nov 22 06:01:05 crc kubenswrapper[4818]: I1122 06:01:05.846472 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29396521-h4m77" event={"ID":"be3d9b50-d430-4a39-85e9-3afb138cfdf4","Type":"ContainerDied","Data":"a45ed19a0c9a6f1cab029647f675d9156f5723ea63dd07a4ac4aecd8e9992509"} Nov 22 06:01:05 crc kubenswrapper[4818]: I1122 06:01:05.903683 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qz69t/must-gather-qq2wp"] Nov 22 06:01:05 crc kubenswrapper[4818]: I1122 06:01:05.905637 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qz69t/must-gather-qq2wp" Nov 22 06:01:05 crc kubenswrapper[4818]: I1122 06:01:05.910240 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qz69t"/"kube-root-ca.crt" Nov 22 06:01:05 crc kubenswrapper[4818]: I1122 06:01:05.910497 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qz69t"/"openshift-service-ca.crt" Nov 22 06:01:05 crc kubenswrapper[4818]: I1122 06:01:05.914232 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qz69t/must-gather-qq2wp"] Nov 22 06:01:06 crc kubenswrapper[4818]: I1122 06:01:06.079026 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kltf\" (UniqueName: \"kubernetes.io/projected/33278cc5-4458-4244-9cf1-fdbcd53bdda4-kube-api-access-2kltf\") pod \"must-gather-qq2wp\" (UID: \"33278cc5-4458-4244-9cf1-fdbcd53bdda4\") " pod="openshift-must-gather-qz69t/must-gather-qq2wp" Nov 22 06:01:06 crc kubenswrapper[4818]: I1122 06:01:06.079179 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/33278cc5-4458-4244-9cf1-fdbcd53bdda4-must-gather-output\") pod \"must-gather-qq2wp\" (UID: \"33278cc5-4458-4244-9cf1-fdbcd53bdda4\") " pod="openshift-must-gather-qz69t/must-gather-qq2wp" Nov 22 06:01:06 crc kubenswrapper[4818]: I1122 06:01:06.181677 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kltf\" (UniqueName: \"kubernetes.io/projected/33278cc5-4458-4244-9cf1-fdbcd53bdda4-kube-api-access-2kltf\") pod \"must-gather-qq2wp\" (UID: \"33278cc5-4458-4244-9cf1-fdbcd53bdda4\") " pod="openshift-must-gather-qz69t/must-gather-qq2wp" Nov 22 06:01:06 crc kubenswrapper[4818]: I1122 06:01:06.181830 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/33278cc5-4458-4244-9cf1-fdbcd53bdda4-must-gather-output\") pod \"must-gather-qq2wp\" (UID: \"33278cc5-4458-4244-9cf1-fdbcd53bdda4\") " pod="openshift-must-gather-qz69t/must-gather-qq2wp" Nov 22 06:01:06 crc kubenswrapper[4818]: I1122 06:01:06.182609 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/33278cc5-4458-4244-9cf1-fdbcd53bdda4-must-gather-output\") pod \"must-gather-qq2wp\" (UID: \"33278cc5-4458-4244-9cf1-fdbcd53bdda4\") " pod="openshift-must-gather-qz69t/must-gather-qq2wp" Nov 22 06:01:06 crc kubenswrapper[4818]: I1122 06:01:06.215577 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kltf\" (UniqueName: \"kubernetes.io/projected/33278cc5-4458-4244-9cf1-fdbcd53bdda4-kube-api-access-2kltf\") pod \"must-gather-qq2wp\" (UID: \"33278cc5-4458-4244-9cf1-fdbcd53bdda4\") " pod="openshift-must-gather-qz69t/must-gather-qq2wp" Nov 22 06:01:06 crc kubenswrapper[4818]: I1122 06:01:06.247343 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qz69t/must-gather-qq2wp" Nov 22 06:01:06 crc kubenswrapper[4818]: I1122 06:01:06.815421 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qz69t/must-gather-qq2wp"] Nov 22 06:01:06 crc kubenswrapper[4818]: I1122 06:01:06.855821 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qz69t/must-gather-qq2wp" event={"ID":"33278cc5-4458-4244-9cf1-fdbcd53bdda4","Type":"ContainerStarted","Data":"4a5a977fa4465f6766df622f4c3fc8979d2e813e005ddf563c1b9eb6b94d252d"} Nov 22 06:01:07 crc kubenswrapper[4818]: I1122 06:01:07.182515 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29396521-h4m77" Nov 22 06:01:07 crc kubenswrapper[4818]: I1122 06:01:07.307353 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be3d9b50-d430-4a39-85e9-3afb138cfdf4-config-data\") pod \"be3d9b50-d430-4a39-85e9-3afb138cfdf4\" (UID: \"be3d9b50-d430-4a39-85e9-3afb138cfdf4\") " Nov 22 06:01:07 crc kubenswrapper[4818]: I1122 06:01:07.307636 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be3d9b50-d430-4a39-85e9-3afb138cfdf4-combined-ca-bundle\") pod \"be3d9b50-d430-4a39-85e9-3afb138cfdf4\" (UID: \"be3d9b50-d430-4a39-85e9-3afb138cfdf4\") " Nov 22 06:01:07 crc kubenswrapper[4818]: I1122 06:01:07.307686 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfn76\" (UniqueName: \"kubernetes.io/projected/be3d9b50-d430-4a39-85e9-3afb138cfdf4-kube-api-access-jfn76\") pod \"be3d9b50-d430-4a39-85e9-3afb138cfdf4\" (UID: \"be3d9b50-d430-4a39-85e9-3afb138cfdf4\") " Nov 22 06:01:07 crc kubenswrapper[4818]: I1122 06:01:07.307781 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/be3d9b50-d430-4a39-85e9-3afb138cfdf4-fernet-keys\") pod \"be3d9b50-d430-4a39-85e9-3afb138cfdf4\" (UID: \"be3d9b50-d430-4a39-85e9-3afb138cfdf4\") " Nov 22 06:01:07 crc kubenswrapper[4818]: I1122 06:01:07.312911 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be3d9b50-d430-4a39-85e9-3afb138cfdf4-kube-api-access-jfn76" (OuterVolumeSpecName: "kube-api-access-jfn76") pod "be3d9b50-d430-4a39-85e9-3afb138cfdf4" (UID: "be3d9b50-d430-4a39-85e9-3afb138cfdf4"). InnerVolumeSpecName "kube-api-access-jfn76". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 06:01:07 crc kubenswrapper[4818]: I1122 06:01:07.313230 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be3d9b50-d430-4a39-85e9-3afb138cfdf4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "be3d9b50-d430-4a39-85e9-3afb138cfdf4" (UID: "be3d9b50-d430-4a39-85e9-3afb138cfdf4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 06:01:07 crc kubenswrapper[4818]: I1122 06:01:07.349733 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be3d9b50-d430-4a39-85e9-3afb138cfdf4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be3d9b50-d430-4a39-85e9-3afb138cfdf4" (UID: "be3d9b50-d430-4a39-85e9-3afb138cfdf4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 06:01:07 crc kubenswrapper[4818]: I1122 06:01:07.374781 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be3d9b50-d430-4a39-85e9-3afb138cfdf4-config-data" (OuterVolumeSpecName: "config-data") pod "be3d9b50-d430-4a39-85e9-3afb138cfdf4" (UID: "be3d9b50-d430-4a39-85e9-3afb138cfdf4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 06:01:07 crc kubenswrapper[4818]: I1122 06:01:07.412242 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be3d9b50-d430-4a39-85e9-3afb138cfdf4-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 06:01:07 crc kubenswrapper[4818]: I1122 06:01:07.412316 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be3d9b50-d430-4a39-85e9-3afb138cfdf4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 06:01:07 crc kubenswrapper[4818]: I1122 06:01:07.412336 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfn76\" (UniqueName: \"kubernetes.io/projected/be3d9b50-d430-4a39-85e9-3afb138cfdf4-kube-api-access-jfn76\") on node \"crc\" DevicePath \"\"" Nov 22 06:01:07 crc kubenswrapper[4818]: I1122 06:01:07.412354 4818 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/be3d9b50-d430-4a39-85e9-3afb138cfdf4-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 22 06:01:07 crc kubenswrapper[4818]: I1122 06:01:07.871557 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29396521-h4m77" event={"ID":"be3d9b50-d430-4a39-85e9-3afb138cfdf4","Type":"ContainerDied","Data":"6fb2750d99b14c54a8d424d922c4419f545d415b23372eb5625924c95a0f049d"} Nov 22 06:01:07 crc kubenswrapper[4818]: I1122 06:01:07.871593 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fb2750d99b14c54a8d424d922c4419f545d415b23372eb5625924c95a0f049d" Nov 22 06:01:07 crc kubenswrapper[4818]: I1122 06:01:07.871613 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29396521-h4m77" Nov 22 06:01:13 crc kubenswrapper[4818]: I1122 06:01:13.971679 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qz69t/must-gather-qq2wp" event={"ID":"33278cc5-4458-4244-9cf1-fdbcd53bdda4","Type":"ContainerStarted","Data":"e2f9649b112209878db60859c8f18382ae5b9bf91a3d9413b8445b4093d447ad"} Nov 22 06:01:14 crc kubenswrapper[4818]: I1122 06:01:14.982354 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qz69t/must-gather-qq2wp" event={"ID":"33278cc5-4458-4244-9cf1-fdbcd53bdda4","Type":"ContainerStarted","Data":"46d616995bf4f5b5ca92e0e6239884e567f70f2f978fb6e2a99606069aa0ba4e"} Nov 22 06:01:15 crc kubenswrapper[4818]: I1122 06:01:15.002781 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-qz69t/must-gather-qq2wp" podStartSLOduration=3.793012847 podStartE2EDuration="10.002766359s" podCreationTimestamp="2025-11-22 06:01:05 +0000 UTC" firstStartedPulling="2025-11-22 06:01:06.825038392 +0000 UTC m=+4419.399454919" lastFinishedPulling="2025-11-22 06:01:13.034791904 +0000 UTC m=+4425.609208431" observedRunningTime="2025-11-22 06:01:15.000613401 +0000 UTC m=+4427.575029928" watchObservedRunningTime="2025-11-22 06:01:15.002766359 +0000 UTC m=+4427.577182886" Nov 22 06:01:17 crc kubenswrapper[4818]: I1122 06:01:17.884078 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qz69t/crc-debug-x56kr"] Nov 22 06:01:17 crc kubenswrapper[4818]: E1122 06:01:17.885171 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be3d9b50-d430-4a39-85e9-3afb138cfdf4" containerName="keystone-cron" Nov 22 06:01:17 crc kubenswrapper[4818]: I1122 06:01:17.885196 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="be3d9b50-d430-4a39-85e9-3afb138cfdf4" containerName="keystone-cron" Nov 22 06:01:17 crc kubenswrapper[4818]: I1122 06:01:17.885600 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="be3d9b50-d430-4a39-85e9-3afb138cfdf4" containerName="keystone-cron" Nov 22 06:01:17 crc kubenswrapper[4818]: I1122 06:01:17.886923 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qz69t/crc-debug-x56kr" Nov 22 06:01:17 crc kubenswrapper[4818]: I1122 06:01:17.888916 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-qz69t"/"default-dockercfg-mjgk2" Nov 22 06:01:18 crc kubenswrapper[4818]: I1122 06:01:18.053412 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l74l2\" (UniqueName: \"kubernetes.io/projected/8ee3ed7e-afec-451a-b7ef-7bf30b619dda-kube-api-access-l74l2\") pod \"crc-debug-x56kr\" (UID: \"8ee3ed7e-afec-451a-b7ef-7bf30b619dda\") " pod="openshift-must-gather-qz69t/crc-debug-x56kr" Nov 22 06:01:18 crc kubenswrapper[4818]: I1122 06:01:18.053503 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8ee3ed7e-afec-451a-b7ef-7bf30b619dda-host\") pod \"crc-debug-x56kr\" (UID: \"8ee3ed7e-afec-451a-b7ef-7bf30b619dda\") " pod="openshift-must-gather-qz69t/crc-debug-x56kr" Nov 22 06:01:18 crc kubenswrapper[4818]: I1122 06:01:18.154858 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l74l2\" (UniqueName: \"kubernetes.io/projected/8ee3ed7e-afec-451a-b7ef-7bf30b619dda-kube-api-access-l74l2\") pod \"crc-debug-x56kr\" (UID: \"8ee3ed7e-afec-451a-b7ef-7bf30b619dda\") " pod="openshift-must-gather-qz69t/crc-debug-x56kr" Nov 22 06:01:18 crc kubenswrapper[4818]: I1122 06:01:18.154927 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8ee3ed7e-afec-451a-b7ef-7bf30b619dda-host\") pod \"crc-debug-x56kr\" (UID: \"8ee3ed7e-afec-451a-b7ef-7bf30b619dda\") " pod="openshift-must-gather-qz69t/crc-debug-x56kr" Nov 22 06:01:18 crc kubenswrapper[4818]: I1122 06:01:18.155113 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8ee3ed7e-afec-451a-b7ef-7bf30b619dda-host\") pod \"crc-debug-x56kr\" (UID: \"8ee3ed7e-afec-451a-b7ef-7bf30b619dda\") " pod="openshift-must-gather-qz69t/crc-debug-x56kr" Nov 22 06:01:18 crc kubenswrapper[4818]: I1122 06:01:18.173547 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l74l2\" (UniqueName: \"kubernetes.io/projected/8ee3ed7e-afec-451a-b7ef-7bf30b619dda-kube-api-access-l74l2\") pod \"crc-debug-x56kr\" (UID: \"8ee3ed7e-afec-451a-b7ef-7bf30b619dda\") " pod="openshift-must-gather-qz69t/crc-debug-x56kr" Nov 22 06:01:18 crc kubenswrapper[4818]: I1122 06:01:18.209107 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qz69t/crc-debug-x56kr" Nov 22 06:01:18 crc kubenswrapper[4818]: W1122 06:01:18.256031 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ee3ed7e_afec_451a_b7ef_7bf30b619dda.slice/crio-fc5b3600fdf43d24c70526a36cf23bd42a32e773b0e64f6707f1460d7115fffb WatchSource:0}: Error finding container fc5b3600fdf43d24c70526a36cf23bd42a32e773b0e64f6707f1460d7115fffb: Status 404 returned error can't find the container with id fc5b3600fdf43d24c70526a36cf23bd42a32e773b0e64f6707f1460d7115fffb Nov 22 06:01:19 crc kubenswrapper[4818]: I1122 06:01:19.028064 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qz69t/crc-debug-x56kr" event={"ID":"8ee3ed7e-afec-451a-b7ef-7bf30b619dda","Type":"ContainerStarted","Data":"fc5b3600fdf43d24c70526a36cf23bd42a32e773b0e64f6707f1460d7115fffb"} Nov 22 06:01:21 crc kubenswrapper[4818]: I1122 06:01:21.265202 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 06:01:21 crc kubenswrapper[4818]: I1122 06:01:21.265705 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 06:01:21 crc kubenswrapper[4818]: I1122 06:01:21.265778 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 06:01:21 crc kubenswrapper[4818]: I1122 06:01:21.266775 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861"} pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 06:01:21 crc kubenswrapper[4818]: I1122 06:01:21.266861 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" containerID="cri-o://0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" gracePeriod=600 Nov 22 06:01:21 crc kubenswrapper[4818]: E1122 06:01:21.412141 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:01:22 crc kubenswrapper[4818]: I1122 06:01:22.059463 4818 generic.go:334] "Generic (PLEG): container finished" podID="0971c9ec-29ce-44eb-9875-ad4664143305" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" exitCode=0 Nov 22 06:01:22 crc kubenswrapper[4818]: I1122 06:01:22.059511 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerDied","Data":"0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861"} Nov 22 06:01:22 crc kubenswrapper[4818]: I1122 06:01:22.059547 4818 scope.go:117] "RemoveContainer" containerID="87e53488d73d6090d237f383bc00e58356c78446d810b06ded7d6522cfcda586" Nov 22 06:01:22 crc kubenswrapper[4818]: I1122 06:01:22.060216 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:01:22 crc kubenswrapper[4818]: E1122 06:01:22.060617 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:01:30 crc kubenswrapper[4818]: I1122 06:01:30.139871 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qz69t/crc-debug-x56kr" event={"ID":"8ee3ed7e-afec-451a-b7ef-7bf30b619dda","Type":"ContainerStarted","Data":"1b2ae5bbe4d125ca94ae39875d74157be068beb854394f99f99a639ff35e19e1"} Nov 22 06:01:30 crc kubenswrapper[4818]: I1122 06:01:30.160021 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-qz69t/crc-debug-x56kr" podStartSLOduration=2.248626075 podStartE2EDuration="13.159996958s" podCreationTimestamp="2025-11-22 06:01:17 +0000 UTC" firstStartedPulling="2025-11-22 06:01:18.257950475 +0000 UTC m=+4430.832367002" lastFinishedPulling="2025-11-22 06:01:29.169321348 +0000 UTC m=+4441.743737885" observedRunningTime="2025-11-22 06:01:30.155085798 +0000 UTC m=+4442.729502325" watchObservedRunningTime="2025-11-22 06:01:30.159996958 +0000 UTC m=+4442.734413505" Nov 22 06:01:32 crc kubenswrapper[4818]: I1122 06:01:32.291340 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:01:32 crc kubenswrapper[4818]: E1122 06:01:32.292106 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:01:43 crc kubenswrapper[4818]: I1122 06:01:43.292180 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:01:43 crc kubenswrapper[4818]: E1122 06:01:43.292871 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:01:55 crc kubenswrapper[4818]: I1122 06:01:55.292636 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:01:55 crc kubenswrapper[4818]: E1122 06:01:55.293510 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:02:07 crc kubenswrapper[4818]: I1122 06:02:07.292513 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:02:07 crc kubenswrapper[4818]: E1122 06:02:07.293356 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:02:19 crc kubenswrapper[4818]: I1122 06:02:19.292405 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:02:19 crc kubenswrapper[4818]: E1122 06:02:19.293612 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:02:21 crc kubenswrapper[4818]: I1122 06:02:21.649558 4818 generic.go:334] "Generic (PLEG): container finished" podID="8ee3ed7e-afec-451a-b7ef-7bf30b619dda" containerID="1b2ae5bbe4d125ca94ae39875d74157be068beb854394f99f99a639ff35e19e1" exitCode=0 Nov 22 06:02:21 crc kubenswrapper[4818]: I1122 06:02:21.650344 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qz69t/crc-debug-x56kr" event={"ID":"8ee3ed7e-afec-451a-b7ef-7bf30b619dda","Type":"ContainerDied","Data":"1b2ae5bbe4d125ca94ae39875d74157be068beb854394f99f99a639ff35e19e1"} Nov 22 06:02:22 crc kubenswrapper[4818]: I1122 06:02:22.775788 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qz69t/crc-debug-x56kr" Nov 22 06:02:22 crc kubenswrapper[4818]: I1122 06:02:22.820390 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qz69t/crc-debug-x56kr"] Nov 22 06:02:22 crc kubenswrapper[4818]: I1122 06:02:22.827981 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qz69t/crc-debug-x56kr"] Nov 22 06:02:22 crc kubenswrapper[4818]: I1122 06:02:22.849794 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l74l2\" (UniqueName: \"kubernetes.io/projected/8ee3ed7e-afec-451a-b7ef-7bf30b619dda-kube-api-access-l74l2\") pod \"8ee3ed7e-afec-451a-b7ef-7bf30b619dda\" (UID: \"8ee3ed7e-afec-451a-b7ef-7bf30b619dda\") " Nov 22 06:02:22 crc kubenswrapper[4818]: I1122 06:02:22.849905 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8ee3ed7e-afec-451a-b7ef-7bf30b619dda-host\") pod \"8ee3ed7e-afec-451a-b7ef-7bf30b619dda\" (UID: \"8ee3ed7e-afec-451a-b7ef-7bf30b619dda\") " Nov 22 06:02:22 crc kubenswrapper[4818]: I1122 06:02:22.850871 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ee3ed7e-afec-451a-b7ef-7bf30b619dda-host" (OuterVolumeSpecName: "host") pod "8ee3ed7e-afec-451a-b7ef-7bf30b619dda" (UID: "8ee3ed7e-afec-451a-b7ef-7bf30b619dda"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 06:02:22 crc kubenswrapper[4818]: I1122 06:02:22.862662 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ee3ed7e-afec-451a-b7ef-7bf30b619dda-kube-api-access-l74l2" (OuterVolumeSpecName: "kube-api-access-l74l2") pod "8ee3ed7e-afec-451a-b7ef-7bf30b619dda" (UID: "8ee3ed7e-afec-451a-b7ef-7bf30b619dda"). InnerVolumeSpecName "kube-api-access-l74l2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 06:02:22 crc kubenswrapper[4818]: I1122 06:02:22.953349 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l74l2\" (UniqueName: \"kubernetes.io/projected/8ee3ed7e-afec-451a-b7ef-7bf30b619dda-kube-api-access-l74l2\") on node \"crc\" DevicePath \"\"" Nov 22 06:02:22 crc kubenswrapper[4818]: I1122 06:02:22.953387 4818 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8ee3ed7e-afec-451a-b7ef-7bf30b619dda-host\") on node \"crc\" DevicePath \"\"" Nov 22 06:02:23 crc kubenswrapper[4818]: I1122 06:02:23.667384 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc5b3600fdf43d24c70526a36cf23bd42a32e773b0e64f6707f1460d7115fffb" Nov 22 06:02:23 crc kubenswrapper[4818]: I1122 06:02:23.667462 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qz69t/crc-debug-x56kr" Nov 22 06:02:24 crc kubenswrapper[4818]: I1122 06:02:24.002022 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qz69t/crc-debug-pj6r7"] Nov 22 06:02:24 crc kubenswrapper[4818]: E1122 06:02:24.002658 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ee3ed7e-afec-451a-b7ef-7bf30b619dda" containerName="container-00" Nov 22 06:02:24 crc kubenswrapper[4818]: I1122 06:02:24.002669 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ee3ed7e-afec-451a-b7ef-7bf30b619dda" containerName="container-00" Nov 22 06:02:24 crc kubenswrapper[4818]: I1122 06:02:24.002833 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ee3ed7e-afec-451a-b7ef-7bf30b619dda" containerName="container-00" Nov 22 06:02:24 crc kubenswrapper[4818]: I1122 06:02:24.003405 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qz69t/crc-debug-pj6r7" Nov 22 06:02:24 crc kubenswrapper[4818]: I1122 06:02:24.004945 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-qz69t"/"default-dockercfg-mjgk2" Nov 22 06:02:24 crc kubenswrapper[4818]: I1122 06:02:24.181542 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgzlb\" (UniqueName: \"kubernetes.io/projected/cc9b89c4-a2c6-404b-93cd-75b6430659f9-kube-api-access-pgzlb\") pod \"crc-debug-pj6r7\" (UID: \"cc9b89c4-a2c6-404b-93cd-75b6430659f9\") " pod="openshift-must-gather-qz69t/crc-debug-pj6r7" Nov 22 06:02:24 crc kubenswrapper[4818]: I1122 06:02:24.181824 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc9b89c4-a2c6-404b-93cd-75b6430659f9-host\") pod \"crc-debug-pj6r7\" (UID: \"cc9b89c4-a2c6-404b-93cd-75b6430659f9\") " pod="openshift-must-gather-qz69t/crc-debug-pj6r7" Nov 22 06:02:24 crc kubenswrapper[4818]: I1122 06:02:24.283670 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc9b89c4-a2c6-404b-93cd-75b6430659f9-host\") pod \"crc-debug-pj6r7\" (UID: \"cc9b89c4-a2c6-404b-93cd-75b6430659f9\") " pod="openshift-must-gather-qz69t/crc-debug-pj6r7" Nov 22 06:02:24 crc kubenswrapper[4818]: I1122 06:02:24.283812 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc9b89c4-a2c6-404b-93cd-75b6430659f9-host\") pod \"crc-debug-pj6r7\" (UID: \"cc9b89c4-a2c6-404b-93cd-75b6430659f9\") " pod="openshift-must-gather-qz69t/crc-debug-pj6r7" Nov 22 06:02:24 crc kubenswrapper[4818]: I1122 06:02:24.283865 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgzlb\" (UniqueName: \"kubernetes.io/projected/cc9b89c4-a2c6-404b-93cd-75b6430659f9-kube-api-access-pgzlb\") pod \"crc-debug-pj6r7\" (UID: \"cc9b89c4-a2c6-404b-93cd-75b6430659f9\") " pod="openshift-must-gather-qz69t/crc-debug-pj6r7" Nov 22 06:02:24 crc kubenswrapper[4818]: I1122 06:02:24.308410 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ee3ed7e-afec-451a-b7ef-7bf30b619dda" path="/var/lib/kubelet/pods/8ee3ed7e-afec-451a-b7ef-7bf30b619dda/volumes" Nov 22 06:02:24 crc kubenswrapper[4818]: I1122 06:02:24.315906 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgzlb\" (UniqueName: \"kubernetes.io/projected/cc9b89c4-a2c6-404b-93cd-75b6430659f9-kube-api-access-pgzlb\") pod \"crc-debug-pj6r7\" (UID: \"cc9b89c4-a2c6-404b-93cd-75b6430659f9\") " pod="openshift-must-gather-qz69t/crc-debug-pj6r7" Nov 22 06:02:24 crc kubenswrapper[4818]: I1122 06:02:24.331599 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qz69t/crc-debug-pj6r7" Nov 22 06:02:24 crc kubenswrapper[4818]: I1122 06:02:24.676803 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qz69t/crc-debug-pj6r7" event={"ID":"cc9b89c4-a2c6-404b-93cd-75b6430659f9","Type":"ContainerStarted","Data":"a8e19c4e29b6f5e1bc4dd38e8525ebbb156f594ec527d104c6369385724a5926"} Nov 22 06:02:24 crc kubenswrapper[4818]: I1122 06:02:24.677136 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qz69t/crc-debug-pj6r7" event={"ID":"cc9b89c4-a2c6-404b-93cd-75b6430659f9","Type":"ContainerStarted","Data":"6e8c0c81cb380045a910d9797da936837f0714e49aed5d3b8a62d16d073d90f9"} Nov 22 06:02:24 crc kubenswrapper[4818]: I1122 06:02:24.698958 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-qz69t/crc-debug-pj6r7" podStartSLOduration=1.698940511 podStartE2EDuration="1.698940511s" podCreationTimestamp="2025-11-22 06:02:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 06:02:24.690896836 +0000 UTC m=+4497.265313403" watchObservedRunningTime="2025-11-22 06:02:24.698940511 +0000 UTC m=+4497.273357038" Nov 22 06:02:25 crc kubenswrapper[4818]: I1122 06:02:25.685761 4818 generic.go:334] "Generic (PLEG): container finished" podID="cc9b89c4-a2c6-404b-93cd-75b6430659f9" containerID="a8e19c4e29b6f5e1bc4dd38e8525ebbb156f594ec527d104c6369385724a5926" exitCode=0 Nov 22 06:02:25 crc kubenswrapper[4818]: I1122 06:02:25.685816 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qz69t/crc-debug-pj6r7" event={"ID":"cc9b89c4-a2c6-404b-93cd-75b6430659f9","Type":"ContainerDied","Data":"a8e19c4e29b6f5e1bc4dd38e8525ebbb156f594ec527d104c6369385724a5926"} Nov 22 06:02:27 crc kubenswrapper[4818]: I1122 06:02:27.421015 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qz69t/crc-debug-pj6r7" Nov 22 06:02:27 crc kubenswrapper[4818]: I1122 06:02:27.442634 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgzlb\" (UniqueName: \"kubernetes.io/projected/cc9b89c4-a2c6-404b-93cd-75b6430659f9-kube-api-access-pgzlb\") pod \"cc9b89c4-a2c6-404b-93cd-75b6430659f9\" (UID: \"cc9b89c4-a2c6-404b-93cd-75b6430659f9\") " Nov 22 06:02:27 crc kubenswrapper[4818]: I1122 06:02:27.442751 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc9b89c4-a2c6-404b-93cd-75b6430659f9-host\") pod \"cc9b89c4-a2c6-404b-93cd-75b6430659f9\" (UID: \"cc9b89c4-a2c6-404b-93cd-75b6430659f9\") " Nov 22 06:02:27 crc kubenswrapper[4818]: I1122 06:02:27.442868 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc9b89c4-a2c6-404b-93cd-75b6430659f9-host" (OuterVolumeSpecName: "host") pod "cc9b89c4-a2c6-404b-93cd-75b6430659f9" (UID: "cc9b89c4-a2c6-404b-93cd-75b6430659f9"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 06:02:27 crc kubenswrapper[4818]: I1122 06:02:27.443714 4818 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc9b89c4-a2c6-404b-93cd-75b6430659f9-host\") on node \"crc\" DevicePath \"\"" Nov 22 06:02:27 crc kubenswrapper[4818]: I1122 06:02:27.449482 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc9b89c4-a2c6-404b-93cd-75b6430659f9-kube-api-access-pgzlb" (OuterVolumeSpecName: "kube-api-access-pgzlb") pod "cc9b89c4-a2c6-404b-93cd-75b6430659f9" (UID: "cc9b89c4-a2c6-404b-93cd-75b6430659f9"). InnerVolumeSpecName "kube-api-access-pgzlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 06:02:27 crc kubenswrapper[4818]: I1122 06:02:27.463777 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qz69t/crc-debug-pj6r7"] Nov 22 06:02:27 crc kubenswrapper[4818]: I1122 06:02:27.471367 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qz69t/crc-debug-pj6r7"] Nov 22 06:02:27 crc kubenswrapper[4818]: I1122 06:02:27.545078 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgzlb\" (UniqueName: \"kubernetes.io/projected/cc9b89c4-a2c6-404b-93cd-75b6430659f9-kube-api-access-pgzlb\") on node \"crc\" DevicePath \"\"" Nov 22 06:02:27 crc kubenswrapper[4818]: I1122 06:02:27.722613 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e8c0c81cb380045a910d9797da936837f0714e49aed5d3b8a62d16d073d90f9" Nov 22 06:02:27 crc kubenswrapper[4818]: I1122 06:02:27.722672 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qz69t/crc-debug-pj6r7" Nov 22 06:02:28 crc kubenswrapper[4818]: I1122 06:02:28.326717 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc9b89c4-a2c6-404b-93cd-75b6430659f9" path="/var/lib/kubelet/pods/cc9b89c4-a2c6-404b-93cd-75b6430659f9/volumes" Nov 22 06:02:28 crc kubenswrapper[4818]: I1122 06:02:28.793239 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qz69t/crc-debug-vv5w6"] Nov 22 06:02:28 crc kubenswrapper[4818]: E1122 06:02:28.793650 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc9b89c4-a2c6-404b-93cd-75b6430659f9" containerName="container-00" Nov 22 06:02:28 crc kubenswrapper[4818]: I1122 06:02:28.793662 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc9b89c4-a2c6-404b-93cd-75b6430659f9" containerName="container-00" Nov 22 06:02:28 crc kubenswrapper[4818]: I1122 06:02:28.793867 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc9b89c4-a2c6-404b-93cd-75b6430659f9" containerName="container-00" Nov 22 06:02:28 crc kubenswrapper[4818]: I1122 06:02:28.794487 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qz69t/crc-debug-vv5w6" Nov 22 06:02:28 crc kubenswrapper[4818]: I1122 06:02:28.796651 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-qz69t"/"default-dockercfg-mjgk2" Nov 22 06:02:28 crc kubenswrapper[4818]: I1122 06:02:28.972927 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmtv6\" (UniqueName: \"kubernetes.io/projected/c0d4aa13-4add-4f31-b959-bbdb0f6c33bf-kube-api-access-lmtv6\") pod \"crc-debug-vv5w6\" (UID: \"c0d4aa13-4add-4f31-b959-bbdb0f6c33bf\") " pod="openshift-must-gather-qz69t/crc-debug-vv5w6" Nov 22 06:02:28 crc kubenswrapper[4818]: I1122 06:02:28.973218 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c0d4aa13-4add-4f31-b959-bbdb0f6c33bf-host\") pod \"crc-debug-vv5w6\" (UID: \"c0d4aa13-4add-4f31-b959-bbdb0f6c33bf\") " pod="openshift-must-gather-qz69t/crc-debug-vv5w6" Nov 22 06:02:29 crc kubenswrapper[4818]: I1122 06:02:29.074865 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmtv6\" (UniqueName: \"kubernetes.io/projected/c0d4aa13-4add-4f31-b959-bbdb0f6c33bf-kube-api-access-lmtv6\") pod \"crc-debug-vv5w6\" (UID: \"c0d4aa13-4add-4f31-b959-bbdb0f6c33bf\") " pod="openshift-must-gather-qz69t/crc-debug-vv5w6" Nov 22 06:02:29 crc kubenswrapper[4818]: I1122 06:02:29.075317 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c0d4aa13-4add-4f31-b959-bbdb0f6c33bf-host\") pod \"crc-debug-vv5w6\" (UID: \"c0d4aa13-4add-4f31-b959-bbdb0f6c33bf\") " pod="openshift-must-gather-qz69t/crc-debug-vv5w6" Nov 22 06:02:29 crc kubenswrapper[4818]: I1122 06:02:29.075438 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c0d4aa13-4add-4f31-b959-bbdb0f6c33bf-host\") pod \"crc-debug-vv5w6\" (UID: \"c0d4aa13-4add-4f31-b959-bbdb0f6c33bf\") " pod="openshift-must-gather-qz69t/crc-debug-vv5w6" Nov 22 06:02:29 crc kubenswrapper[4818]: I1122 06:02:29.099994 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmtv6\" (UniqueName: \"kubernetes.io/projected/c0d4aa13-4add-4f31-b959-bbdb0f6c33bf-kube-api-access-lmtv6\") pod \"crc-debug-vv5w6\" (UID: \"c0d4aa13-4add-4f31-b959-bbdb0f6c33bf\") " pod="openshift-must-gather-qz69t/crc-debug-vv5w6" Nov 22 06:02:29 crc kubenswrapper[4818]: I1122 06:02:29.112897 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qz69t/crc-debug-vv5w6" Nov 22 06:02:29 crc kubenswrapper[4818]: W1122 06:02:29.169443 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0d4aa13_4add_4f31_b959_bbdb0f6c33bf.slice/crio-25c2468335b9229b5217200cba5e3396331776deaf8090cd98efb2e1e6bf09cc WatchSource:0}: Error finding container 25c2468335b9229b5217200cba5e3396331776deaf8090cd98efb2e1e6bf09cc: Status 404 returned error can't find the container with id 25c2468335b9229b5217200cba5e3396331776deaf8090cd98efb2e1e6bf09cc Nov 22 06:02:29 crc kubenswrapper[4818]: I1122 06:02:29.746614 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qz69t/crc-debug-vv5w6" event={"ID":"c0d4aa13-4add-4f31-b959-bbdb0f6c33bf","Type":"ContainerStarted","Data":"81a039e8a5926b0f2a3b660d204481d0c8c6e0828d0d2bd96f6092052ff2d2d2"} Nov 22 06:02:29 crc kubenswrapper[4818]: I1122 06:02:29.747006 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qz69t/crc-debug-vv5w6" event={"ID":"c0d4aa13-4add-4f31-b959-bbdb0f6c33bf","Type":"ContainerStarted","Data":"25c2468335b9229b5217200cba5e3396331776deaf8090cd98efb2e1e6bf09cc"} Nov 22 06:02:30 crc kubenswrapper[4818]: I1122 06:02:30.764619 4818 generic.go:334] "Generic (PLEG): container finished" podID="c0d4aa13-4add-4f31-b959-bbdb0f6c33bf" containerID="81a039e8a5926b0f2a3b660d204481d0c8c6e0828d0d2bd96f6092052ff2d2d2" exitCode=0 Nov 22 06:02:30 crc kubenswrapper[4818]: I1122 06:02:30.764729 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qz69t/crc-debug-vv5w6" event={"ID":"c0d4aa13-4add-4f31-b959-bbdb0f6c33bf","Type":"ContainerDied","Data":"81a039e8a5926b0f2a3b660d204481d0c8c6e0828d0d2bd96f6092052ff2d2d2"} Nov 22 06:02:30 crc kubenswrapper[4818]: I1122 06:02:30.819645 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qz69t/crc-debug-vv5w6"] Nov 22 06:02:30 crc kubenswrapper[4818]: I1122 06:02:30.830023 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qz69t/crc-debug-vv5w6"] Nov 22 06:02:31 crc kubenswrapper[4818]: I1122 06:02:31.864916 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qz69t/crc-debug-vv5w6" Nov 22 06:02:32 crc kubenswrapper[4818]: I1122 06:02:32.040143 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c0d4aa13-4add-4f31-b959-bbdb0f6c33bf-host\") pod \"c0d4aa13-4add-4f31-b959-bbdb0f6c33bf\" (UID: \"c0d4aa13-4add-4f31-b959-bbdb0f6c33bf\") " Nov 22 06:02:32 crc kubenswrapper[4818]: I1122 06:02:32.040237 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmtv6\" (UniqueName: \"kubernetes.io/projected/c0d4aa13-4add-4f31-b959-bbdb0f6c33bf-kube-api-access-lmtv6\") pod \"c0d4aa13-4add-4f31-b959-bbdb0f6c33bf\" (UID: \"c0d4aa13-4add-4f31-b959-bbdb0f6c33bf\") " Nov 22 06:02:32 crc kubenswrapper[4818]: I1122 06:02:32.040279 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c0d4aa13-4add-4f31-b959-bbdb0f6c33bf-host" (OuterVolumeSpecName: "host") pod "c0d4aa13-4add-4f31-b959-bbdb0f6c33bf" (UID: "c0d4aa13-4add-4f31-b959-bbdb0f6c33bf"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 06:02:32 crc kubenswrapper[4818]: I1122 06:02:32.041087 4818 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c0d4aa13-4add-4f31-b959-bbdb0f6c33bf-host\") on node \"crc\" DevicePath \"\"" Nov 22 06:02:32 crc kubenswrapper[4818]: I1122 06:02:32.046950 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0d4aa13-4add-4f31-b959-bbdb0f6c33bf-kube-api-access-lmtv6" (OuterVolumeSpecName: "kube-api-access-lmtv6") pod "c0d4aa13-4add-4f31-b959-bbdb0f6c33bf" (UID: "c0d4aa13-4add-4f31-b959-bbdb0f6c33bf"). InnerVolumeSpecName "kube-api-access-lmtv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 06:02:32 crc kubenswrapper[4818]: I1122 06:02:32.142963 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmtv6\" (UniqueName: \"kubernetes.io/projected/c0d4aa13-4add-4f31-b959-bbdb0f6c33bf-kube-api-access-lmtv6\") on node \"crc\" DevicePath \"\"" Nov 22 06:02:32 crc kubenswrapper[4818]: I1122 06:02:32.303944 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0d4aa13-4add-4f31-b959-bbdb0f6c33bf" path="/var/lib/kubelet/pods/c0d4aa13-4add-4f31-b959-bbdb0f6c33bf/volumes" Nov 22 06:02:32 crc kubenswrapper[4818]: I1122 06:02:32.787533 4818 scope.go:117] "RemoveContainer" containerID="81a039e8a5926b0f2a3b660d204481d0c8c6e0828d0d2bd96f6092052ff2d2d2" Nov 22 06:02:32 crc kubenswrapper[4818]: I1122 06:02:32.787600 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qz69t/crc-debug-vv5w6" Nov 22 06:02:33 crc kubenswrapper[4818]: I1122 06:02:33.291470 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:02:33 crc kubenswrapper[4818]: E1122 06:02:33.292045 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:02:44 crc kubenswrapper[4818]: I1122 06:02:44.291980 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:02:44 crc kubenswrapper[4818]: E1122 06:02:44.292959 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:02:56 crc kubenswrapper[4818]: I1122 06:02:56.291370 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:02:56 crc kubenswrapper[4818]: E1122 06:02:56.292065 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:03:03 crc kubenswrapper[4818]: I1122 06:03:03.878152 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6dfb48dcf4-6ph2v_1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d/barbican-api/0.log" Nov 22 06:03:03 crc kubenswrapper[4818]: I1122 06:03:03.929165 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6dfb48dcf4-6ph2v_1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d/barbican-api-log/0.log" Nov 22 06:03:04 crc kubenswrapper[4818]: I1122 06:03:04.082326 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5c65499f7d-s6w7z_a5925d65-97cd-493a-8114-6ed07de4386d/barbican-keystone-listener/0.log" Nov 22 06:03:04 crc kubenswrapper[4818]: I1122 06:03:04.299828 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5c65499f7d-s6w7z_a5925d65-97cd-493a-8114-6ed07de4386d/barbican-keystone-listener-log/0.log" Nov 22 06:03:04 crc kubenswrapper[4818]: I1122 06:03:04.321514 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-fb9947b9-nbdwz_3958bab3-ce66-4d2d-8a76-75e1b7593199/barbican-worker/0.log" Nov 22 06:03:04 crc kubenswrapper[4818]: I1122 06:03:04.325960 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-fb9947b9-nbdwz_3958bab3-ce66-4d2d-8a76-75e1b7593199/barbican-worker-log/0.log" Nov 22 06:03:04 crc kubenswrapper[4818]: I1122 06:03:04.493385 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g_597bc073-9fa3-485c-b23d-8f7b5ea21da9/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:03:04 crc kubenswrapper[4818]: I1122 06:03:04.579554 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4ea3ac76-3418-4b1c-8800-d57f554e50e3/ceilometer-central-agent/0.log" Nov 22 06:03:04 crc kubenswrapper[4818]: I1122 06:03:04.689687 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4ea3ac76-3418-4b1c-8800-d57f554e50e3/ceilometer-notification-agent/0.log" Nov 22 06:03:04 crc kubenswrapper[4818]: I1122 06:03:04.708666 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4ea3ac76-3418-4b1c-8800-d57f554e50e3/proxy-httpd/0.log" Nov 22 06:03:04 crc kubenswrapper[4818]: I1122 06:03:04.767990 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4ea3ac76-3418-4b1c-8800-d57f554e50e3/sg-core/0.log" Nov 22 06:03:04 crc kubenswrapper[4818]: I1122 06:03:04.877820 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2_1488ee6c-884b-4a61-93c9-ae582c9f4778/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:03:04 crc kubenswrapper[4818]: I1122 06:03:04.961886 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8_5fe8b800-5e58-48ae-8da3-eae014028909/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:03:05 crc kubenswrapper[4818]: I1122 06:03:05.128985 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9a4f90e5-006b-43db-b910-429b496a7c65/cinder-api-log/0.log" Nov 22 06:03:05 crc kubenswrapper[4818]: I1122 06:03:05.155236 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9a4f90e5-006b-43db-b910-429b496a7c65/cinder-api/0.log" Nov 22 06:03:05 crc kubenswrapper[4818]: I1122 06:03:05.377432 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_0c1e7078-787f-4725-80b1-45f7047e108f/cinder-backup/0.log" Nov 22 06:03:05 crc kubenswrapper[4818]: I1122 06:03:05.399387 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_0c1e7078-787f-4725-80b1-45f7047e108f/probe/0.log" Nov 22 06:03:05 crc kubenswrapper[4818]: I1122 06:03:05.466295 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_0fce1ca4-c9be-4e2a-947a-3cc652edeb50/cinder-scheduler/0.log" Nov 22 06:03:05 crc kubenswrapper[4818]: I1122 06:03:05.652717 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_0fce1ca4-c9be-4e2a-947a-3cc652edeb50/probe/0.log" Nov 22 06:03:05 crc kubenswrapper[4818]: I1122 06:03:05.770710 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_3dfef980-4eab-42b9-821e-233a264843cf/probe/0.log" Nov 22 06:03:05 crc kubenswrapper[4818]: I1122 06:03:05.783802 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_3dfef980-4eab-42b9-821e-233a264843cf/cinder-volume/0.log" Nov 22 06:03:05 crc kubenswrapper[4818]: I1122 06:03:05.861181 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6_4e6cd425-5f7c-4f17-864b-2e3d62851c58/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:03:06 crc kubenswrapper[4818]: I1122 06:03:06.133450 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z_d3f48375-4d0b-4876-b4e8-64a9fa970b63/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:03:06 crc kubenswrapper[4818]: I1122 06:03:06.190817 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-zl4gj_cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b/init/0.log" Nov 22 06:03:06 crc kubenswrapper[4818]: I1122 06:03:06.325282 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-zl4gj_cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b/init/0.log" Nov 22 06:03:06 crc kubenswrapper[4818]: I1122 06:03:06.372918 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-zl4gj_cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b/dnsmasq-dns/0.log" Nov 22 06:03:06 crc kubenswrapper[4818]: I1122 06:03:06.386386 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_6652494e-6fe5-4c2b-b265-bb2e5567265e/glance-httpd/0.log" Nov 22 06:03:06 crc kubenswrapper[4818]: I1122 06:03:06.512059 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_6652494e-6fe5-4c2b-b265-bb2e5567265e/glance-log/0.log" Nov 22 06:03:06 crc kubenswrapper[4818]: I1122 06:03:06.603607 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f87508a5-9796-46dd-b9a4-29815fd608ad/glance-httpd/0.log" Nov 22 06:03:06 crc kubenswrapper[4818]: I1122 06:03:06.634243 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f87508a5-9796-46dd-b9a4-29815fd608ad/glance-log/0.log" Nov 22 06:03:06 crc kubenswrapper[4818]: I1122 06:03:06.889713 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-86757bfcb-s9bbk_34f2acb8-906b-4520-86c2-bb603a964489/horizon/0.log" Nov 22 06:03:06 crc kubenswrapper[4818]: I1122 06:03:06.892811 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-6h722_69bfd3ff-b747-43c0-82ab-ced23f3bc502/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:03:06 crc kubenswrapper[4818]: I1122 06:03:06.937892 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-86757bfcb-s9bbk_34f2acb8-906b-4520-86c2-bb603a964489/horizon-log/0.log" Nov 22 06:03:07 crc kubenswrapper[4818]: I1122 06:03:07.027384 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-fld7g_969421c2-6440-41e6-9866-c624dfe1cbae/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:03:07 crc kubenswrapper[4818]: I1122 06:03:07.245360 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29396521-h4m77_be3d9b50-d430-4a39-85e9-3afb138cfdf4/keystone-cron/0.log" Nov 22 06:03:07 crc kubenswrapper[4818]: I1122 06:03:07.338998 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_fd030ac8-aebe-4cde-b0e8-403f3049a922/kube-state-metrics/0.log" Nov 22 06:03:07 crc kubenswrapper[4818]: I1122 06:03:07.526447 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-595lb_806c9298-1aad-4e84-a80a-b2078f548d03/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:03:07 crc kubenswrapper[4818]: I1122 06:03:07.911902 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_7feb3390-f51a-43f4-86c6-f44c30ebe2ab/manila-api/0.log" Nov 22 06:03:07 crc kubenswrapper[4818]: I1122 06:03:07.977070 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a/probe/0.log" Nov 22 06:03:08 crc kubenswrapper[4818]: I1122 06:03:08.092102 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a/manila-scheduler/0.log" Nov 22 06:03:08 crc kubenswrapper[4818]: I1122 06:03:08.282645 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_a5573538-84fb-4f11-9186-0dfd22a2232c/probe/0.log" Nov 22 06:03:08 crc kubenswrapper[4818]: I1122 06:03:08.312897 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-54d7c687fd-74zvh_fe562192-8025-4265-826d-6e7b1400b281/keystone-api/0.log" Nov 22 06:03:08 crc kubenswrapper[4818]: I1122 06:03:08.474618 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_a5573538-84fb-4f11-9186-0dfd22a2232c/manila-share/0.log" Nov 22 06:03:08 crc kubenswrapper[4818]: I1122 06:03:08.553705 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_7feb3390-f51a-43f4-86c6-f44c30ebe2ab/manila-api-log/0.log" Nov 22 06:03:08 crc kubenswrapper[4818]: I1122 06:03:08.662789 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-64754b546f-kqpj9_81f8a1cb-f2b5-429b-b12b-d2ff49b2e888/neutron-httpd/0.log" Nov 22 06:03:08 crc kubenswrapper[4818]: I1122 06:03:08.675631 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-64754b546f-kqpj9_81f8a1cb-f2b5-429b-b12b-d2ff49b2e888/neutron-api/0.log" Nov 22 06:03:08 crc kubenswrapper[4818]: I1122 06:03:08.837752 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z_e80824ae-495b-475f-a266-44879bb04e25/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:03:09 crc kubenswrapper[4818]: I1122 06:03:09.144579 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_ed42de57-a99e-4940-9874-49f6f703f8aa/nova-api-log/0.log" Nov 22 06:03:09 crc kubenswrapper[4818]: I1122 06:03:09.266970 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_5f15bdce-2f90-40df-a66e-141fb5d9f79f/nova-cell0-conductor-conductor/0.log" Nov 22 06:03:09 crc kubenswrapper[4818]: I1122 06:03:09.431385 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_ed42de57-a99e-4940-9874-49f6f703f8aa/nova-api-api/0.log" Nov 22 06:03:09 crc kubenswrapper[4818]: I1122 06:03:09.479187 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_f9d0ba71-b442-4c8a-87bd-72a8ca7284b2/nova-cell1-conductor-conductor/0.log" Nov 22 06:03:09 crc kubenswrapper[4818]: I1122 06:03:09.638875 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_3f5dcaf3-117e-4024-b1fc-41878a0eefd8/nova-cell1-novncproxy-novncproxy/0.log" Nov 22 06:03:09 crc kubenswrapper[4818]: I1122 06:03:09.774040 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q_38964063-2a2d-4728-9ccc-d5eca9db132d/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:03:09 crc kubenswrapper[4818]: I1122 06:03:09.983474 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e2b602e6-0bbf-4137-a806-d974b00003d7/nova-metadata-log/0.log" Nov 22 06:03:10 crc kubenswrapper[4818]: I1122 06:03:10.183850 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_f9271e30-e243-4016-bc57-2a10e514deb0/nova-scheduler-scheduler/0.log" Nov 22 06:03:10 crc kubenswrapper[4818]: I1122 06:03:10.224147 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_72cac26c-ba82-41c7-a405-a9162867abea/mysql-bootstrap/0.log" Nov 22 06:03:10 crc kubenswrapper[4818]: I1122 06:03:10.292639 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:03:10 crc kubenswrapper[4818]: E1122 06:03:10.292878 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:03:11 crc kubenswrapper[4818]: I1122 06:03:11.010911 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_72cac26c-ba82-41c7-a405-a9162867abea/mysql-bootstrap/0.log" Nov 22 06:03:11 crc kubenswrapper[4818]: I1122 06:03:11.083762 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_72cac26c-ba82-41c7-a405-a9162867abea/galera/0.log" Nov 22 06:03:11 crc kubenswrapper[4818]: I1122 06:03:11.217351 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_4714282b-39ba-4691-8fbd-c3e76120fbb5/mysql-bootstrap/0.log" Nov 22 06:03:11 crc kubenswrapper[4818]: I1122 06:03:11.396915 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_4714282b-39ba-4691-8fbd-c3e76120fbb5/mysql-bootstrap/0.log" Nov 22 06:03:11 crc kubenswrapper[4818]: I1122 06:03:11.432785 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_4714282b-39ba-4691-8fbd-c3e76120fbb5/galera/0.log" Nov 22 06:03:11 crc kubenswrapper[4818]: I1122 06:03:11.565847 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_ac27ad40-d8e9-4ebb-9623-3136a1f9753c/openstackclient/0.log" Nov 22 06:03:11 crc kubenswrapper[4818]: I1122 06:03:11.592562 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e2b602e6-0bbf-4137-a806-d974b00003d7/nova-metadata-metadata/0.log" Nov 22 06:03:11 crc kubenswrapper[4818]: I1122 06:03:11.652536 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-jgq5d_7dec9ecf-e438-4334-a20e-757ce6fff445/ovn-controller/0.log" Nov 22 06:03:12 crc kubenswrapper[4818]: I1122 06:03:12.321178 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-rsg7p_5b90b727-2867-4825-a5c1-a3549da28378/openstack-network-exporter/0.log" Nov 22 06:03:12 crc kubenswrapper[4818]: I1122 06:03:12.342727 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kpdh5_1b31a0d3-3489-430a-853a-40d8715f7e7f/ovsdb-server-init/0.log" Nov 22 06:03:12 crc kubenswrapper[4818]: I1122 06:03:12.514758 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kpdh5_1b31a0d3-3489-430a-853a-40d8715f7e7f/ovs-vswitchd/0.log" Nov 22 06:03:12 crc kubenswrapper[4818]: I1122 06:03:12.558335 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kpdh5_1b31a0d3-3489-430a-853a-40d8715f7e7f/ovsdb-server/0.log" Nov 22 06:03:12 crc kubenswrapper[4818]: I1122 06:03:12.597184 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kpdh5_1b31a0d3-3489-430a-853a-40d8715f7e7f/ovsdb-server-init/0.log" Nov 22 06:03:12 crc kubenswrapper[4818]: I1122 06:03:12.730202 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-k2rlm_0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:03:12 crc kubenswrapper[4818]: I1122 06:03:12.832527 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_6ab8b0e2-6df1-4a6b-b800-93e5b51ee430/openstack-network-exporter/0.log" Nov 22 06:03:12 crc kubenswrapper[4818]: I1122 06:03:12.870731 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_6ab8b0e2-6df1-4a6b-b800-93e5b51ee430/ovn-northd/0.log" Nov 22 06:03:13 crc kubenswrapper[4818]: I1122 06:03:13.034678 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_52d9e522-f84d-4671-9981-d08b3d7e0bbf/openstack-network-exporter/0.log" Nov 22 06:03:13 crc kubenswrapper[4818]: I1122 06:03:13.061571 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_52d9e522-f84d-4671-9981-d08b3d7e0bbf/ovsdbserver-nb/0.log" Nov 22 06:03:13 crc kubenswrapper[4818]: I1122 06:03:13.179518 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_5b0746b7-8400-44c8-b4c5-c6b6e3811e91/openstack-network-exporter/0.log" Nov 22 06:03:13 crc kubenswrapper[4818]: I1122 06:03:13.244566 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_5b0746b7-8400-44c8-b4c5-c6b6e3811e91/ovsdbserver-sb/0.log" Nov 22 06:03:13 crc kubenswrapper[4818]: I1122 06:03:13.447754 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7b464767cb-jcwc8_4e48ac54-5e91-4747-b9e9-28aeb21180fd/placement-api/0.log" Nov 22 06:03:13 crc kubenswrapper[4818]: I1122 06:03:13.590071 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7b464767cb-jcwc8_4e48ac54-5e91-4747-b9e9-28aeb21180fd/placement-log/0.log" Nov 22 06:03:13 crc kubenswrapper[4818]: I1122 06:03:13.625666 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3e1d70d4-6ada-47bf-a7f0-086ac6b891a0/setup-container/0.log" Nov 22 06:03:13 crc kubenswrapper[4818]: I1122 06:03:13.862701 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3e1d70d4-6ada-47bf-a7f0-086ac6b891a0/setup-container/0.log" Nov 22 06:03:13 crc kubenswrapper[4818]: I1122 06:03:13.942763 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3e1d70d4-6ada-47bf-a7f0-086ac6b891a0/rabbitmq/0.log" Nov 22 06:03:13 crc kubenswrapper[4818]: I1122 06:03:13.945483 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86/setup-container/0.log" Nov 22 06:03:14 crc kubenswrapper[4818]: I1122 06:03:14.166477 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86/setup-container/0.log" Nov 22 06:03:14 crc kubenswrapper[4818]: I1122 06:03:14.180235 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx_a072b598-ba0f-4b0d-8978-2d6e824ff5f6/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:03:14 crc kubenswrapper[4818]: I1122 06:03:14.215560 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86/rabbitmq/0.log" Nov 22 06:03:14 crc kubenswrapper[4818]: I1122 06:03:14.355185 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst_0c466daa-c1c7-4d3c-a502-cfb8f424f6b2/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:03:14 crc kubenswrapper[4818]: I1122 06:03:14.445122 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-8b757_5d2cfc6e-9af5-4110-add7-b03c44ffbd33/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:03:14 crc kubenswrapper[4818]: I1122 06:03:14.638093 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-bg6f2_376ee003-816b-45fa-a57d-b9d2fa8120eb/ssh-known-hosts-edpm-deployment/0.log" Nov 22 06:03:14 crc kubenswrapper[4818]: I1122 06:03:14.713412 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_df264e71-c1de-4e1f-9a56-24f5bd284a80/tempest-tests-tempest-tests-runner/0.log" Nov 22 06:03:14 crc kubenswrapper[4818]: I1122 06:03:14.829910 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_fce70419-9c81-461c-bcb5-e829c1868e03/test-operator-logs-container/0.log" Nov 22 06:03:14 crc kubenswrapper[4818]: I1122 06:03:14.983813 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx_68cb051d-9556-47e4-acc2-3bdb9b60ebdd/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:03:23 crc kubenswrapper[4818]: I1122 06:03:23.291004 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:03:23 crc kubenswrapper[4818]: E1122 06:03:23.291681 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:03:32 crc kubenswrapper[4818]: I1122 06:03:32.122184 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_e98d9180-5e2b-4b38-8fc0-844072603c2d/memcached/0.log" Nov 22 06:03:38 crc kubenswrapper[4818]: I1122 06:03:38.297921 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:03:38 crc kubenswrapper[4818]: E1122 06:03:38.299295 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:03:42 crc kubenswrapper[4818]: I1122 06:03:42.027744 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-8vknn_69877edd-8b97-4838-bc41-c087fc4deb7e/kube-rbac-proxy/0.log" Nov 22 06:03:42 crc kubenswrapper[4818]: I1122 06:03:42.078678 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-8vknn_69877edd-8b97-4838-bc41-c087fc4deb7e/manager/0.log" Nov 22 06:03:42 crc kubenswrapper[4818]: I1122 06:03:42.214845 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-b2lp5_27ec8b4f-e5ec-483a-ada8-6a766b843b73/kube-rbac-proxy/0.log" Nov 22 06:03:42 crc kubenswrapper[4818]: I1122 06:03:42.288214 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-b2lp5_27ec8b4f-e5ec-483a-ada8-6a766b843b73/manager/0.log" Nov 22 06:03:42 crc kubenswrapper[4818]: I1122 06:03:42.390634 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-jttdm_b9a85b40-dd76-45b2-bd41-ecb5be7387da/kube-rbac-proxy/0.log" Nov 22 06:03:42 crc kubenswrapper[4818]: I1122 06:03:42.393717 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-jttdm_b9a85b40-dd76-45b2-bd41-ecb5be7387da/manager/0.log" Nov 22 06:03:42 crc kubenswrapper[4818]: I1122 06:03:42.492396 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw_a5e6bbd8-bd14-4785-addd-67c8028cbd22/util/0.log" Nov 22 06:03:42 crc kubenswrapper[4818]: I1122 06:03:42.698565 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw_a5e6bbd8-bd14-4785-addd-67c8028cbd22/util/0.log" Nov 22 06:03:42 crc kubenswrapper[4818]: I1122 06:03:42.742749 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw_a5e6bbd8-bd14-4785-addd-67c8028cbd22/pull/0.log" Nov 22 06:03:42 crc kubenswrapper[4818]: I1122 06:03:42.753242 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw_a5e6bbd8-bd14-4785-addd-67c8028cbd22/pull/0.log" Nov 22 06:03:42 crc kubenswrapper[4818]: I1122 06:03:42.920739 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw_a5e6bbd8-bd14-4785-addd-67c8028cbd22/util/0.log" Nov 22 06:03:42 crc kubenswrapper[4818]: I1122 06:03:42.933151 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw_a5e6bbd8-bd14-4785-addd-67c8028cbd22/extract/0.log" Nov 22 06:03:42 crc kubenswrapper[4818]: I1122 06:03:42.947789 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw_a5e6bbd8-bd14-4785-addd-67c8028cbd22/pull/0.log" Nov 22 06:03:43 crc kubenswrapper[4818]: I1122 06:03:43.085735 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-cl4tl_4d915797-cd52-4f07-9028-4f80411fe65b/kube-rbac-proxy/0.log" Nov 22 06:03:43 crc kubenswrapper[4818]: I1122 06:03:43.168239 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-qt9nt_10da5c26-c5b3-4477-a4d4-a5c98741dfb1/kube-rbac-proxy/0.log" Nov 22 06:03:43 crc kubenswrapper[4818]: I1122 06:03:43.179589 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-cl4tl_4d915797-cd52-4f07-9028-4f80411fe65b/manager/0.log" Nov 22 06:03:43 crc kubenswrapper[4818]: I1122 06:03:43.292424 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-qt9nt_10da5c26-c5b3-4477-a4d4-a5c98741dfb1/manager/0.log" Nov 22 06:03:43 crc kubenswrapper[4818]: I1122 06:03:43.369476 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-4zl42_05bab737-6cda-4891-a014-202ab05b141c/kube-rbac-proxy/0.log" Nov 22 06:03:43 crc kubenswrapper[4818]: I1122 06:03:43.402751 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-4zl42_05bab737-6cda-4891-a014-202ab05b141c/manager/0.log" Nov 22 06:03:43 crc kubenswrapper[4818]: I1122 06:03:43.551592 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7875d8bb94-dn8xx_7389ff62-006d-4d11-8237-e57c9279b0f3/kube-rbac-proxy/0.log" Nov 22 06:03:43 crc kubenswrapper[4818]: I1122 06:03:43.707910 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7875d8bb94-dn8xx_7389ff62-006d-4d11-8237-e57c9279b0f3/manager/0.log" Nov 22 06:03:43 crc kubenswrapper[4818]: I1122 06:03:43.745842 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-ncfxv_c31258b3-6cac-41e8-a350-1ea7b360b848/kube-rbac-proxy/0.log" Nov 22 06:03:43 crc kubenswrapper[4818]: I1122 06:03:43.752076 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-ncfxv_c31258b3-6cac-41e8-a350-1ea7b360b848/manager/0.log" Nov 22 06:03:43 crc kubenswrapper[4818]: I1122 06:03:43.908527 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-pqd4c_d6b59804-efc5-4f60-b9d3-bf0e998c8f89/kube-rbac-proxy/0.log" Nov 22 06:03:44 crc kubenswrapper[4818]: I1122 06:03:44.002464 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-pqd4c_d6b59804-efc5-4f60-b9d3-bf0e998c8f89/manager/0.log" Nov 22 06:03:44 crc kubenswrapper[4818]: I1122 06:03:44.032268 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-85bf8d6cb7-lzmmp_4838407d-e782-4aab-8fbf-861f3ba9019a/kube-rbac-proxy/0.log" Nov 22 06:03:44 crc kubenswrapper[4818]: I1122 06:03:44.154068 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-85bf8d6cb7-lzmmp_4838407d-e782-4aab-8fbf-861f3ba9019a/manager/0.log" Nov 22 06:03:44 crc kubenswrapper[4818]: I1122 06:03:44.177083 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-8wk6c_96a59329-8f96-43a5-9abc-fa8d1d7009fb/kube-rbac-proxy/0.log" Nov 22 06:03:44 crc kubenswrapper[4818]: I1122 06:03:44.250196 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-8wk6c_96a59329-8f96-43a5-9abc-fa8d1d7009fb/manager/0.log" Nov 22 06:03:44 crc kubenswrapper[4818]: I1122 06:03:44.351966 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-xtplb_68668b7f-5fc3-4774-8f0d-3eae71839d52/kube-rbac-proxy/0.log" Nov 22 06:03:44 crc kubenswrapper[4818]: I1122 06:03:44.427063 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-xtplb_68668b7f-5fc3-4774-8f0d-3eae71839d52/manager/0.log" Nov 22 06:03:44 crc kubenswrapper[4818]: I1122 06:03:44.534821 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-lx77w_531548a2-b8d3-4a2a-a6e9-1bb2af1891d8/kube-rbac-proxy/0.log" Nov 22 06:03:44 crc kubenswrapper[4818]: I1122 06:03:44.629822 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-lx77w_531548a2-b8d3-4a2a-a6e9-1bb2af1891d8/manager/0.log" Nov 22 06:03:44 crc kubenswrapper[4818]: I1122 06:03:44.701973 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-rf9vq_10d7f6df-ae52-4dab-b49d-764d95f77935/kube-rbac-proxy/0.log" Nov 22 06:03:44 crc kubenswrapper[4818]: I1122 06:03:44.730529 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-rf9vq_10d7f6df-ae52-4dab-b49d-764d95f77935/manager/0.log" Nov 22 06:03:44 crc kubenswrapper[4818]: I1122 06:03:44.884446 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-n54fq_9797ea62-91f5-4b4f-a9f0-27911cf629e4/kube-rbac-proxy/0.log" Nov 22 06:03:44 crc kubenswrapper[4818]: I1122 06:03:44.907727 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-n54fq_9797ea62-91f5-4b4f-a9f0-27911cf629e4/manager/0.log" Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.080883 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6b57b64d88-pnqbd_44f3f34b-8fa2-479c-a876-947fab621708/kube-rbac-proxy/0.log" Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.160668 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-59b954bb6d-6chhn_04e2d18d-98fc-4fa8-9a1f-1742f497d737/kube-rbac-proxy/0.log" Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.246362 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-krhxl"] Nov 22 06:03:45 crc kubenswrapper[4818]: E1122 06:03:45.246754 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0d4aa13-4add-4f31-b959-bbdb0f6c33bf" containerName="container-00" Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.246770 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0d4aa13-4add-4f31-b959-bbdb0f6c33bf" containerName="container-00" Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.246957 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0d4aa13-4add-4f31-b959-bbdb0f6c33bf" containerName="container-00" Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.249004 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krhxl" Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.260965 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztkvp\" (UniqueName: \"kubernetes.io/projected/56766bb1-3788-4f94-aa85-55694c74c73f-kube-api-access-ztkvp\") pod \"redhat-operators-krhxl\" (UID: \"56766bb1-3788-4f94-aa85-55694c74c73f\") " pod="openshift-marketplace/redhat-operators-krhxl" Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.261028 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56766bb1-3788-4f94-aa85-55694c74c73f-catalog-content\") pod \"redhat-operators-krhxl\" (UID: \"56766bb1-3788-4f94-aa85-55694c74c73f\") " pod="openshift-marketplace/redhat-operators-krhxl" Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.261064 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56766bb1-3788-4f94-aa85-55694c74c73f-utilities\") pod \"redhat-operators-krhxl\" (UID: \"56766bb1-3788-4f94-aa85-55694c74c73f\") " pod="openshift-marketplace/redhat-operators-krhxl" Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.289485 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-krhxl"] Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.362637 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztkvp\" (UniqueName: \"kubernetes.io/projected/56766bb1-3788-4f94-aa85-55694c74c73f-kube-api-access-ztkvp\") pod \"redhat-operators-krhxl\" (UID: \"56766bb1-3788-4f94-aa85-55694c74c73f\") " pod="openshift-marketplace/redhat-operators-krhxl" Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.362958 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56766bb1-3788-4f94-aa85-55694c74c73f-catalog-content\") pod \"redhat-operators-krhxl\" (UID: \"56766bb1-3788-4f94-aa85-55694c74c73f\") " pod="openshift-marketplace/redhat-operators-krhxl" Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.363061 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56766bb1-3788-4f94-aa85-55694c74c73f-utilities\") pod \"redhat-operators-krhxl\" (UID: \"56766bb1-3788-4f94-aa85-55694c74c73f\") " pod="openshift-marketplace/redhat-operators-krhxl" Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.363691 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56766bb1-3788-4f94-aa85-55694c74c73f-utilities\") pod \"redhat-operators-krhxl\" (UID: \"56766bb1-3788-4f94-aa85-55694c74c73f\") " pod="openshift-marketplace/redhat-operators-krhxl" Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.364030 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56766bb1-3788-4f94-aa85-55694c74c73f-catalog-content\") pod \"redhat-operators-krhxl\" (UID: \"56766bb1-3788-4f94-aa85-55694c74c73f\") " pod="openshift-marketplace/redhat-operators-krhxl" Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.435018 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztkvp\" (UniqueName: \"kubernetes.io/projected/56766bb1-3788-4f94-aa85-55694c74c73f-kube-api-access-ztkvp\") pod \"redhat-operators-krhxl\" (UID: \"56766bb1-3788-4f94-aa85-55694c74c73f\") " pod="openshift-marketplace/redhat-operators-krhxl" Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.507013 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-tn592_124c736f-9e78-422d-b317-b7263a45c55e/registry-server/0.log" Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.577123 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krhxl" Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.735064 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-59b954bb6d-6chhn_04e2d18d-98fc-4fa8-9a1f-1742f497d737/operator/0.log" Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.787167 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-8dhvf_06ec2990-cd45-4808-9748-2a1d75c5015f/kube-rbac-proxy/0.log" Nov 22 06:03:45 crc kubenswrapper[4818]: I1122 06:03:45.934213 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-8dhvf_06ec2990-cd45-4808-9748-2a1d75c5015f/manager/0.log" Nov 22 06:03:46 crc kubenswrapper[4818]: I1122 06:03:46.065144 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-56phb_786ab73e-012b-47f6-b033-9fa9aab6d6cb/kube-rbac-proxy/0.log" Nov 22 06:03:46 crc kubenswrapper[4818]: I1122 06:03:46.142004 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-56phb_786ab73e-012b-47f6-b033-9fa9aab6d6cb/manager/0.log" Nov 22 06:03:46 crc kubenswrapper[4818]: I1122 06:03:46.190720 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-krhxl"] Nov 22 06:03:46 crc kubenswrapper[4818]: I1122 06:03:46.324623 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-d6957_62c4e1cf-72fb-4a9b-88a1-350c1e20f360/operator/0.log" Nov 22 06:03:46 crc kubenswrapper[4818]: I1122 06:03:46.370243 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-vkj22_f04a965b-f5f4-45af-843f-862498956853/kube-rbac-proxy/0.log" Nov 22 06:03:46 crc kubenswrapper[4818]: I1122 06:03:46.387974 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6b57b64d88-pnqbd_44f3f34b-8fa2-479c-a876-947fab621708/manager/0.log" Nov 22 06:03:46 crc kubenswrapper[4818]: I1122 06:03:46.467302 4818 generic.go:334] "Generic (PLEG): container finished" podID="56766bb1-3788-4f94-aa85-55694c74c73f" containerID="de526f5e0f280d880206c0387c31b888c5dc29af7fe175372a83f9a50beb18f4" exitCode=0 Nov 22 06:03:46 crc kubenswrapper[4818]: I1122 06:03:46.467353 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krhxl" event={"ID":"56766bb1-3788-4f94-aa85-55694c74c73f","Type":"ContainerDied","Data":"de526f5e0f280d880206c0387c31b888c5dc29af7fe175372a83f9a50beb18f4"} Nov 22 06:03:46 crc kubenswrapper[4818]: I1122 06:03:46.467616 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krhxl" event={"ID":"56766bb1-3788-4f94-aa85-55694c74c73f","Type":"ContainerStarted","Data":"1c1ef85aea3aa48977cfb940c4ee8d8ec36c343df950d1767ffbea4057b16ddb"} Nov 22 06:03:46 crc kubenswrapper[4818]: I1122 06:03:46.645784 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6d4bf84b58-cjphz_bba055bd-511b-4af6-bc94-4cadf7a84853/kube-rbac-proxy/0.log" Nov 22 06:03:46 crc kubenswrapper[4818]: I1122 06:03:46.687549 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-vkj22_f04a965b-f5f4-45af-843f-862498956853/manager/0.log" Nov 22 06:03:46 crc kubenswrapper[4818]: I1122 06:03:46.758902 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6d4bf84b58-cjphz_bba055bd-511b-4af6-bc94-4cadf7a84853/manager/0.log" Nov 22 06:03:46 crc kubenswrapper[4818]: I1122 06:03:46.890744 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-pb45n_e342ff02-d6d9-4be0-b97f-87dea6653a61/manager/0.log" Nov 22 06:03:46 crc kubenswrapper[4818]: I1122 06:03:46.927247 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-pb45n_e342ff02-d6d9-4be0-b97f-87dea6653a61/kube-rbac-proxy/0.log" Nov 22 06:03:46 crc kubenswrapper[4818]: I1122 06:03:46.954794 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-sqvdq_2b80bb62-ec21-44c6-91e2-b422b9813255/kube-rbac-proxy/0.log" Nov 22 06:03:46 crc kubenswrapper[4818]: I1122 06:03:46.968814 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-sqvdq_2b80bb62-ec21-44c6-91e2-b422b9813255/manager/0.log" Nov 22 06:03:47 crc kubenswrapper[4818]: I1122 06:03:47.479938 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krhxl" event={"ID":"56766bb1-3788-4f94-aa85-55694c74c73f","Type":"ContainerStarted","Data":"ea91a681fad3b785f303a53aa8ddd6443c394a517b1877c17529cc1fdcf56f95"} Nov 22 06:03:48 crc kubenswrapper[4818]: I1122 06:03:48.497586 4818 generic.go:334] "Generic (PLEG): container finished" podID="56766bb1-3788-4f94-aa85-55694c74c73f" containerID="ea91a681fad3b785f303a53aa8ddd6443c394a517b1877c17529cc1fdcf56f95" exitCode=0 Nov 22 06:03:48 crc kubenswrapper[4818]: I1122 06:03:48.497638 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krhxl" event={"ID":"56766bb1-3788-4f94-aa85-55694c74c73f","Type":"ContainerDied","Data":"ea91a681fad3b785f303a53aa8ddd6443c394a517b1877c17529cc1fdcf56f95"} Nov 22 06:03:49 crc kubenswrapper[4818]: I1122 06:03:49.507680 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krhxl" event={"ID":"56766bb1-3788-4f94-aa85-55694c74c73f","Type":"ContainerStarted","Data":"1745b394060c01c7da82d8b3b8fc0fd2752fed9b06bad895683985ae6471817c"} Nov 22 06:03:49 crc kubenswrapper[4818]: I1122 06:03:49.530064 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-krhxl" podStartSLOduration=2.077440868 podStartE2EDuration="4.530045417s" podCreationTimestamp="2025-11-22 06:03:45 +0000 UTC" firstStartedPulling="2025-11-22 06:03:46.469703835 +0000 UTC m=+4579.044120362" lastFinishedPulling="2025-11-22 06:03:48.922308374 +0000 UTC m=+4581.496724911" observedRunningTime="2025-11-22 06:03:49.524518259 +0000 UTC m=+4582.098934786" watchObservedRunningTime="2025-11-22 06:03:49.530045417 +0000 UTC m=+4582.104461944" Nov 22 06:03:50 crc kubenswrapper[4818]: I1122 06:03:50.390075 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:03:50 crc kubenswrapper[4818]: E1122 06:03:50.390673 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:03:55 crc kubenswrapper[4818]: I1122 06:03:55.578430 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-krhxl" Nov 22 06:03:55 crc kubenswrapper[4818]: I1122 06:03:55.579044 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-krhxl" Nov 22 06:03:55 crc kubenswrapper[4818]: I1122 06:03:55.632331 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-krhxl" Nov 22 06:03:56 crc kubenswrapper[4818]: I1122 06:03:56.644583 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-krhxl" Nov 22 06:03:56 crc kubenswrapper[4818]: I1122 06:03:56.691907 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-krhxl"] Nov 22 06:03:58 crc kubenswrapper[4818]: I1122 06:03:58.590479 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-krhxl" podUID="56766bb1-3788-4f94-aa85-55694c74c73f" containerName="registry-server" containerID="cri-o://1745b394060c01c7da82d8b3b8fc0fd2752fed9b06bad895683985ae6471817c" gracePeriod=2 Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.031555 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krhxl" Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.175834 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztkvp\" (UniqueName: \"kubernetes.io/projected/56766bb1-3788-4f94-aa85-55694c74c73f-kube-api-access-ztkvp\") pod \"56766bb1-3788-4f94-aa85-55694c74c73f\" (UID: \"56766bb1-3788-4f94-aa85-55694c74c73f\") " Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.175916 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56766bb1-3788-4f94-aa85-55694c74c73f-utilities\") pod \"56766bb1-3788-4f94-aa85-55694c74c73f\" (UID: \"56766bb1-3788-4f94-aa85-55694c74c73f\") " Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.176083 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56766bb1-3788-4f94-aa85-55694c74c73f-catalog-content\") pod \"56766bb1-3788-4f94-aa85-55694c74c73f\" (UID: \"56766bb1-3788-4f94-aa85-55694c74c73f\") " Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.177145 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56766bb1-3788-4f94-aa85-55694c74c73f-utilities" (OuterVolumeSpecName: "utilities") pod "56766bb1-3788-4f94-aa85-55694c74c73f" (UID: "56766bb1-3788-4f94-aa85-55694c74c73f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.180584 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56766bb1-3788-4f94-aa85-55694c74c73f-kube-api-access-ztkvp" (OuterVolumeSpecName: "kube-api-access-ztkvp") pod "56766bb1-3788-4f94-aa85-55694c74c73f" (UID: "56766bb1-3788-4f94-aa85-55694c74c73f"). InnerVolumeSpecName "kube-api-access-ztkvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.198243 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztkvp\" (UniqueName: \"kubernetes.io/projected/56766bb1-3788-4f94-aa85-55694c74c73f-kube-api-access-ztkvp\") on node \"crc\" DevicePath \"\"" Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.198304 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56766bb1-3788-4f94-aa85-55694c74c73f-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.261029 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56766bb1-3788-4f94-aa85-55694c74c73f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56766bb1-3788-4f94-aa85-55694c74c73f" (UID: "56766bb1-3788-4f94-aa85-55694c74c73f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.299787 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56766bb1-3788-4f94-aa85-55694c74c73f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.610582 4818 generic.go:334] "Generic (PLEG): container finished" podID="56766bb1-3788-4f94-aa85-55694c74c73f" containerID="1745b394060c01c7da82d8b3b8fc0fd2752fed9b06bad895683985ae6471817c" exitCode=0 Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.610652 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krhxl" Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.610669 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krhxl" event={"ID":"56766bb1-3788-4f94-aa85-55694c74c73f","Type":"ContainerDied","Data":"1745b394060c01c7da82d8b3b8fc0fd2752fed9b06bad895683985ae6471817c"} Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.611075 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krhxl" event={"ID":"56766bb1-3788-4f94-aa85-55694c74c73f","Type":"ContainerDied","Data":"1c1ef85aea3aa48977cfb940c4ee8d8ec36c343df950d1767ffbea4057b16ddb"} Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.611096 4818 scope.go:117] "RemoveContainer" containerID="1745b394060c01c7da82d8b3b8fc0fd2752fed9b06bad895683985ae6471817c" Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.630206 4818 scope.go:117] "RemoveContainer" containerID="ea91a681fad3b785f303a53aa8ddd6443c394a517b1877c17529cc1fdcf56f95" Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.659944 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-krhxl"] Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.667851 4818 scope.go:117] "RemoveContainer" containerID="de526f5e0f280d880206c0387c31b888c5dc29af7fe175372a83f9a50beb18f4" Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.668942 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-krhxl"] Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.707098 4818 scope.go:117] "RemoveContainer" containerID="1745b394060c01c7da82d8b3b8fc0fd2752fed9b06bad895683985ae6471817c" Nov 22 06:03:59 crc kubenswrapper[4818]: E1122 06:03:59.707447 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1745b394060c01c7da82d8b3b8fc0fd2752fed9b06bad895683985ae6471817c\": container with ID starting with 1745b394060c01c7da82d8b3b8fc0fd2752fed9b06bad895683985ae6471817c not found: ID does not exist" containerID="1745b394060c01c7da82d8b3b8fc0fd2752fed9b06bad895683985ae6471817c" Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.707477 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1745b394060c01c7da82d8b3b8fc0fd2752fed9b06bad895683985ae6471817c"} err="failed to get container status \"1745b394060c01c7da82d8b3b8fc0fd2752fed9b06bad895683985ae6471817c\": rpc error: code = NotFound desc = could not find container \"1745b394060c01c7da82d8b3b8fc0fd2752fed9b06bad895683985ae6471817c\": container with ID starting with 1745b394060c01c7da82d8b3b8fc0fd2752fed9b06bad895683985ae6471817c not found: ID does not exist" Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.707499 4818 scope.go:117] "RemoveContainer" containerID="ea91a681fad3b785f303a53aa8ddd6443c394a517b1877c17529cc1fdcf56f95" Nov 22 06:03:59 crc kubenswrapper[4818]: E1122 06:03:59.707816 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea91a681fad3b785f303a53aa8ddd6443c394a517b1877c17529cc1fdcf56f95\": container with ID starting with ea91a681fad3b785f303a53aa8ddd6443c394a517b1877c17529cc1fdcf56f95 not found: ID does not exist" containerID="ea91a681fad3b785f303a53aa8ddd6443c394a517b1877c17529cc1fdcf56f95" Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.707840 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea91a681fad3b785f303a53aa8ddd6443c394a517b1877c17529cc1fdcf56f95"} err="failed to get container status \"ea91a681fad3b785f303a53aa8ddd6443c394a517b1877c17529cc1fdcf56f95\": rpc error: code = NotFound desc = could not find container \"ea91a681fad3b785f303a53aa8ddd6443c394a517b1877c17529cc1fdcf56f95\": container with ID starting with ea91a681fad3b785f303a53aa8ddd6443c394a517b1877c17529cc1fdcf56f95 not found: ID does not exist" Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.707853 4818 scope.go:117] "RemoveContainer" containerID="de526f5e0f280d880206c0387c31b888c5dc29af7fe175372a83f9a50beb18f4" Nov 22 06:03:59 crc kubenswrapper[4818]: E1122 06:03:59.708104 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de526f5e0f280d880206c0387c31b888c5dc29af7fe175372a83f9a50beb18f4\": container with ID starting with de526f5e0f280d880206c0387c31b888c5dc29af7fe175372a83f9a50beb18f4 not found: ID does not exist" containerID="de526f5e0f280d880206c0387c31b888c5dc29af7fe175372a83f9a50beb18f4" Nov 22 06:03:59 crc kubenswrapper[4818]: I1122 06:03:59.708157 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de526f5e0f280d880206c0387c31b888c5dc29af7fe175372a83f9a50beb18f4"} err="failed to get container status \"de526f5e0f280d880206c0387c31b888c5dc29af7fe175372a83f9a50beb18f4\": rpc error: code = NotFound desc = could not find container \"de526f5e0f280d880206c0387c31b888c5dc29af7fe175372a83f9a50beb18f4\": container with ID starting with de526f5e0f280d880206c0387c31b888c5dc29af7fe175372a83f9a50beb18f4 not found: ID does not exist" Nov 22 06:04:00 crc kubenswrapper[4818]: I1122 06:04:00.304590 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56766bb1-3788-4f94-aa85-55694c74c73f" path="/var/lib/kubelet/pods/56766bb1-3788-4f94-aa85-55694c74c73f/volumes" Nov 22 06:04:04 crc kubenswrapper[4818]: I1122 06:04:04.291512 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:04:04 crc kubenswrapper[4818]: E1122 06:04:04.292121 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:04:04 crc kubenswrapper[4818]: I1122 06:04:04.655011 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zftfz_7c429c72-1ae2-4b7d-9786-57ed9341bd48/control-plane-machine-set-operator/0.log" Nov 22 06:04:04 crc kubenswrapper[4818]: I1122 06:04:04.784106 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-8kzxl_8d3b850f-8917-46af-b713-891a7c979d2e/kube-rbac-proxy/0.log" Nov 22 06:04:04 crc kubenswrapper[4818]: I1122 06:04:04.826454 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-8kzxl_8d3b850f-8917-46af-b713-891a7c979d2e/machine-api-operator/0.log" Nov 22 06:04:17 crc kubenswrapper[4818]: I1122 06:04:17.291938 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:04:17 crc kubenswrapper[4818]: E1122 06:04:17.292680 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:04:19 crc kubenswrapper[4818]: I1122 06:04:19.716077 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-9l45k_c62c123f-00fc-4a5e-9649-87121c5f793a/cert-manager-controller/0.log" Nov 22 06:04:20 crc kubenswrapper[4818]: I1122 06:04:20.333479 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-248l2_de576ad7-b4e6-4047-92d1-dc19f477be33/cert-manager-cainjector/0.log" Nov 22 06:04:20 crc kubenswrapper[4818]: I1122 06:04:20.352048 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-4bmxf_77efc3b2-d4e5-426e-ada4-de8c900575f9/cert-manager-webhook/0.log" Nov 22 06:04:28 crc kubenswrapper[4818]: I1122 06:04:28.305425 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:04:28 crc kubenswrapper[4818]: E1122 06:04:28.306813 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:04:35 crc kubenswrapper[4818]: I1122 06:04:35.183125 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-d2f5x_614393c8-1594-4694-8281-a1e20cab9209/nmstate-console-plugin/0.log" Nov 22 06:04:35 crc kubenswrapper[4818]: I1122 06:04:35.371629 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-f8v82_38740f9a-864c-44b3-8f95-f9ffecc8e9d5/kube-rbac-proxy/0.log" Nov 22 06:04:35 crc kubenswrapper[4818]: I1122 06:04:35.387449 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-f8v82_38740f9a-864c-44b3-8f95-f9ffecc8e9d5/nmstate-metrics/0.log" Nov 22 06:04:35 crc kubenswrapper[4818]: I1122 06:04:35.390226 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-ftnnk_695d974e-caa1-43c9-b619-363891b2bb7c/nmstate-handler/0.log" Nov 22 06:04:35 crc kubenswrapper[4818]: I1122 06:04:35.581769 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-r2wjf_05f359e3-f133-443a-9067-148ddc6fac3b/nmstate-operator/0.log" Nov 22 06:04:35 crc kubenswrapper[4818]: I1122 06:04:35.591998 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-d94tc_b0f0e1e5-3f4d-463b-8c00-91db5867fdfc/nmstate-webhook/0.log" Nov 22 06:04:39 crc kubenswrapper[4818]: I1122 06:04:39.291608 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:04:39 crc kubenswrapper[4818]: E1122 06:04:39.292287 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:04:50 crc kubenswrapper[4818]: I1122 06:04:50.548131 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-wbzs6_0fdd1082-4473-4604-8ced-2294bf4ba248/kube-rbac-proxy/0.log" Nov 22 06:04:50 crc kubenswrapper[4818]: I1122 06:04:50.668692 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-frr-files/0.log" Nov 22 06:04:50 crc kubenswrapper[4818]: I1122 06:04:50.718429 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-wbzs6_0fdd1082-4473-4604-8ced-2294bf4ba248/controller/0.log" Nov 22 06:04:50 crc kubenswrapper[4818]: I1122 06:04:50.830830 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-reloader/0.log" Nov 22 06:04:50 crc kubenswrapper[4818]: I1122 06:04:50.835562 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-metrics/0.log" Nov 22 06:04:50 crc kubenswrapper[4818]: I1122 06:04:50.836390 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-frr-files/0.log" Nov 22 06:04:50 crc kubenswrapper[4818]: I1122 06:04:50.919575 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-reloader/0.log" Nov 22 06:04:51 crc kubenswrapper[4818]: I1122 06:04:51.095787 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-reloader/0.log" Nov 22 06:04:51 crc kubenswrapper[4818]: I1122 06:04:51.131145 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-metrics/0.log" Nov 22 06:04:51 crc kubenswrapper[4818]: I1122 06:04:51.164698 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-frr-files/0.log" Nov 22 06:04:51 crc kubenswrapper[4818]: I1122 06:04:51.192214 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-metrics/0.log" Nov 22 06:04:51 crc kubenswrapper[4818]: I1122 06:04:51.324211 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-frr-files/0.log" Nov 22 06:04:51 crc kubenswrapper[4818]: I1122 06:04:51.387028 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-metrics/0.log" Nov 22 06:04:51 crc kubenswrapper[4818]: I1122 06:04:51.395687 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/controller/0.log" Nov 22 06:04:51 crc kubenswrapper[4818]: I1122 06:04:51.446378 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-reloader/0.log" Nov 22 06:04:51 crc kubenswrapper[4818]: I1122 06:04:51.570399 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/kube-rbac-proxy/0.log" Nov 22 06:04:51 crc kubenswrapper[4818]: I1122 06:04:51.594646 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/frr-metrics/0.log" Nov 22 06:04:51 crc kubenswrapper[4818]: I1122 06:04:51.701406 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/kube-rbac-proxy-frr/0.log" Nov 22 06:04:51 crc kubenswrapper[4818]: I1122 06:04:51.850311 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/reloader/0.log" Nov 22 06:04:51 crc kubenswrapper[4818]: I1122 06:04:51.952839 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-rss4n_a90765d9-f4ec-4f76-921b-f8fe52fa2cdc/frr-k8s-webhook-server/0.log" Nov 22 06:04:52 crc kubenswrapper[4818]: I1122 06:04:52.060823 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5b7cbb7cc5-ff54p_254a82ef-b456-4309-93fb-a236494bf3d4/manager/0.log" Nov 22 06:04:52 crc kubenswrapper[4818]: I1122 06:04:52.178050 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-757d8876f9-6qjkv_63241779-80e0-43a3-bb9c-6530a4a6ae05/webhook-server/0.log" Nov 22 06:04:52 crc kubenswrapper[4818]: I1122 06:04:52.291103 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:04:52 crc kubenswrapper[4818]: E1122 06:04:52.291431 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:04:52 crc kubenswrapper[4818]: I1122 06:04:52.331887 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8mwxx_dbc8ab54-ce5e-4320-8deb-4d597cb4f82a/kube-rbac-proxy/0.log" Nov 22 06:04:52 crc kubenswrapper[4818]: I1122 06:04:52.779187 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8mwxx_dbc8ab54-ce5e-4320-8deb-4d597cb4f82a/speaker/0.log" Nov 22 06:04:52 crc kubenswrapper[4818]: I1122 06:04:52.917158 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/frr/0.log" Nov 22 06:05:04 crc kubenswrapper[4818]: I1122 06:05:04.291681 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:05:04 crc kubenswrapper[4818]: E1122 06:05:04.292759 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:05:07 crc kubenswrapper[4818]: I1122 06:05:07.320297 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd_cc4bf9a2-88ba-4882-9908-f1e2f27c82ac/util/0.log" Nov 22 06:05:07 crc kubenswrapper[4818]: I1122 06:05:07.496183 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd_cc4bf9a2-88ba-4882-9908-f1e2f27c82ac/util/0.log" Nov 22 06:05:07 crc kubenswrapper[4818]: I1122 06:05:07.505120 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd_cc4bf9a2-88ba-4882-9908-f1e2f27c82ac/pull/0.log" Nov 22 06:05:07 crc kubenswrapper[4818]: I1122 06:05:07.535207 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd_cc4bf9a2-88ba-4882-9908-f1e2f27c82ac/pull/0.log" Nov 22 06:05:07 crc kubenswrapper[4818]: I1122 06:05:07.703910 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd_cc4bf9a2-88ba-4882-9908-f1e2f27c82ac/util/0.log" Nov 22 06:05:07 crc kubenswrapper[4818]: I1122 06:05:07.735029 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd_cc4bf9a2-88ba-4882-9908-f1e2f27c82ac/pull/0.log" Nov 22 06:05:07 crc kubenswrapper[4818]: I1122 06:05:07.742529 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd_cc4bf9a2-88ba-4882-9908-f1e2f27c82ac/extract/0.log" Nov 22 06:05:07 crc kubenswrapper[4818]: I1122 06:05:07.866171 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcnvz_603888a5-b8ab-4378-ab59-d044221e403c/extract-utilities/0.log" Nov 22 06:05:08 crc kubenswrapper[4818]: I1122 06:05:08.028266 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcnvz_603888a5-b8ab-4378-ab59-d044221e403c/extract-content/0.log" Nov 22 06:05:08 crc kubenswrapper[4818]: I1122 06:05:08.043873 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcnvz_603888a5-b8ab-4378-ab59-d044221e403c/extract-utilities/0.log" Nov 22 06:05:08 crc kubenswrapper[4818]: I1122 06:05:08.059106 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcnvz_603888a5-b8ab-4378-ab59-d044221e403c/extract-content/0.log" Nov 22 06:05:08 crc kubenswrapper[4818]: I1122 06:05:08.238877 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcnvz_603888a5-b8ab-4378-ab59-d044221e403c/extract-content/0.log" Nov 22 06:05:08 crc kubenswrapper[4818]: I1122 06:05:08.254870 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcnvz_603888a5-b8ab-4378-ab59-d044221e403c/extract-utilities/0.log" Nov 22 06:05:08 crc kubenswrapper[4818]: I1122 06:05:08.914445 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cdjr2_fd603ec5-ddaf-4d51-a4fc-09442ae6692e/extract-utilities/0.log" Nov 22 06:05:08 crc kubenswrapper[4818]: I1122 06:05:08.946279 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcnvz_603888a5-b8ab-4378-ab59-d044221e403c/registry-server/0.log" Nov 22 06:05:09 crc kubenswrapper[4818]: I1122 06:05:09.063514 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cdjr2_fd603ec5-ddaf-4d51-a4fc-09442ae6692e/extract-content/0.log" Nov 22 06:05:09 crc kubenswrapper[4818]: I1122 06:05:09.121404 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cdjr2_fd603ec5-ddaf-4d51-a4fc-09442ae6692e/extract-content/0.log" Nov 22 06:05:09 crc kubenswrapper[4818]: I1122 06:05:09.131452 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cdjr2_fd603ec5-ddaf-4d51-a4fc-09442ae6692e/extract-utilities/0.log" Nov 22 06:05:09 crc kubenswrapper[4818]: I1122 06:05:09.257028 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cdjr2_fd603ec5-ddaf-4d51-a4fc-09442ae6692e/extract-utilities/0.log" Nov 22 06:05:09 crc kubenswrapper[4818]: I1122 06:05:09.261902 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cdjr2_fd603ec5-ddaf-4d51-a4fc-09442ae6692e/extract-content/0.log" Nov 22 06:05:09 crc kubenswrapper[4818]: I1122 06:05:09.476458 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cdjr2_fd603ec5-ddaf-4d51-a4fc-09442ae6692e/registry-server/0.log" Nov 22 06:05:09 crc kubenswrapper[4818]: I1122 06:05:09.516035 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d_cd479145-6121-48b1-b54a-b122e24a03fd/util/0.log" Nov 22 06:05:09 crc kubenswrapper[4818]: I1122 06:05:09.651129 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d_cd479145-6121-48b1-b54a-b122e24a03fd/util/0.log" Nov 22 06:05:09 crc kubenswrapper[4818]: I1122 06:05:09.651763 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d_cd479145-6121-48b1-b54a-b122e24a03fd/pull/0.log" Nov 22 06:05:09 crc kubenswrapper[4818]: I1122 06:05:09.653211 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d_cd479145-6121-48b1-b54a-b122e24a03fd/pull/0.log" Nov 22 06:05:09 crc kubenswrapper[4818]: I1122 06:05:09.803979 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d_cd479145-6121-48b1-b54a-b122e24a03fd/pull/0.log" Nov 22 06:05:09 crc kubenswrapper[4818]: I1122 06:05:09.838860 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d_cd479145-6121-48b1-b54a-b122e24a03fd/extract/0.log" Nov 22 06:05:09 crc kubenswrapper[4818]: I1122 06:05:09.881833 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d_cd479145-6121-48b1-b54a-b122e24a03fd/util/0.log" Nov 22 06:05:09 crc kubenswrapper[4818]: I1122 06:05:09.938827 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-cxqxn_6e9cbedb-a168-495f-8746-ccf21c4e330c/marketplace-operator/0.log" Nov 22 06:05:10 crc kubenswrapper[4818]: I1122 06:05:10.043083 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rcbjp_d83d4bfc-8842-4d00-96d9-779a89080441/extract-utilities/0.log" Nov 22 06:05:10 crc kubenswrapper[4818]: I1122 06:05:10.198653 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rcbjp_d83d4bfc-8842-4d00-96d9-779a89080441/extract-utilities/0.log" Nov 22 06:05:10 crc kubenswrapper[4818]: I1122 06:05:10.203648 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rcbjp_d83d4bfc-8842-4d00-96d9-779a89080441/extract-content/0.log" Nov 22 06:05:10 crc kubenswrapper[4818]: I1122 06:05:10.218571 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rcbjp_d83d4bfc-8842-4d00-96d9-779a89080441/extract-content/0.log" Nov 22 06:05:10 crc kubenswrapper[4818]: I1122 06:05:10.357740 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rcbjp_d83d4bfc-8842-4d00-96d9-779a89080441/extract-content/0.log" Nov 22 06:05:10 crc kubenswrapper[4818]: I1122 06:05:10.365697 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rcbjp_d83d4bfc-8842-4d00-96d9-779a89080441/extract-utilities/0.log" Nov 22 06:05:10 crc kubenswrapper[4818]: I1122 06:05:10.479435 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8lq9_759a3ee8-f990-4960-bc61-9893ca8cbc6f/extract-utilities/0.log" Nov 22 06:05:10 crc kubenswrapper[4818]: I1122 06:05:10.505941 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rcbjp_d83d4bfc-8842-4d00-96d9-779a89080441/registry-server/0.log" Nov 22 06:05:10 crc kubenswrapper[4818]: I1122 06:05:10.639545 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8lq9_759a3ee8-f990-4960-bc61-9893ca8cbc6f/extract-utilities/0.log" Nov 22 06:05:10 crc kubenswrapper[4818]: I1122 06:05:10.643481 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8lq9_759a3ee8-f990-4960-bc61-9893ca8cbc6f/extract-content/0.log" Nov 22 06:05:10 crc kubenswrapper[4818]: I1122 06:05:10.681189 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8lq9_759a3ee8-f990-4960-bc61-9893ca8cbc6f/extract-content/0.log" Nov 22 06:05:10 crc kubenswrapper[4818]: I1122 06:05:10.776679 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8lq9_759a3ee8-f990-4960-bc61-9893ca8cbc6f/extract-utilities/0.log" Nov 22 06:05:10 crc kubenswrapper[4818]: I1122 06:05:10.787740 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8lq9_759a3ee8-f990-4960-bc61-9893ca8cbc6f/extract-content/0.log" Nov 22 06:05:11 crc kubenswrapper[4818]: I1122 06:05:11.507401 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8lq9_759a3ee8-f990-4960-bc61-9893ca8cbc6f/registry-server/0.log" Nov 22 06:05:15 crc kubenswrapper[4818]: I1122 06:05:15.291734 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:05:15 crc kubenswrapper[4818]: E1122 06:05:15.292608 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:05:27 crc kubenswrapper[4818]: I1122 06:05:27.291656 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:05:27 crc kubenswrapper[4818]: E1122 06:05:27.292327 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:05:42 crc kubenswrapper[4818]: I1122 06:05:42.292096 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:05:42 crc kubenswrapper[4818]: E1122 06:05:42.292734 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:05:56 crc kubenswrapper[4818]: I1122 06:05:56.291235 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:05:56 crc kubenswrapper[4818]: E1122 06:05:56.293829 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:06:08 crc kubenswrapper[4818]: I1122 06:06:08.303384 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:06:08 crc kubenswrapper[4818]: E1122 06:06:08.304016 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:06:19 crc kubenswrapper[4818]: I1122 06:06:19.294105 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:06:19 crc kubenswrapper[4818]: E1122 06:06:19.295061 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:06:33 crc kubenswrapper[4818]: I1122 06:06:33.291631 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:06:34 crc kubenswrapper[4818]: I1122 06:06:34.241988 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"d2100269659d0a3f462da37bed1df37938f3da7494989b3a814f1fdfb9d5a0f9"} Nov 22 06:06:36 crc kubenswrapper[4818]: I1122 06:06:36.669867 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gtmnc"] Nov 22 06:06:36 crc kubenswrapper[4818]: E1122 06:06:36.671041 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56766bb1-3788-4f94-aa85-55694c74c73f" containerName="extract-content" Nov 22 06:06:36 crc kubenswrapper[4818]: I1122 06:06:36.671058 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="56766bb1-3788-4f94-aa85-55694c74c73f" containerName="extract-content" Nov 22 06:06:36 crc kubenswrapper[4818]: E1122 06:06:36.671075 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56766bb1-3788-4f94-aa85-55694c74c73f" containerName="extract-utilities" Nov 22 06:06:36 crc kubenswrapper[4818]: I1122 06:06:36.671084 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="56766bb1-3788-4f94-aa85-55694c74c73f" containerName="extract-utilities" Nov 22 06:06:36 crc kubenswrapper[4818]: E1122 06:06:36.671125 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56766bb1-3788-4f94-aa85-55694c74c73f" containerName="registry-server" Nov 22 06:06:36 crc kubenswrapper[4818]: I1122 06:06:36.671135 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="56766bb1-3788-4f94-aa85-55694c74c73f" containerName="registry-server" Nov 22 06:06:36 crc kubenswrapper[4818]: I1122 06:06:36.671508 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="56766bb1-3788-4f94-aa85-55694c74c73f" containerName="registry-server" Nov 22 06:06:36 crc kubenswrapper[4818]: I1122 06:06:36.673621 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gtmnc" Nov 22 06:06:36 crc kubenswrapper[4818]: I1122 06:06:36.697071 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gtmnc"] Nov 22 06:06:36 crc kubenswrapper[4818]: I1122 06:06:36.753809 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2103acc2-9d59-4f08-97d2-f1cbee955f68-catalog-content\") pod \"redhat-marketplace-gtmnc\" (UID: \"2103acc2-9d59-4f08-97d2-f1cbee955f68\") " pod="openshift-marketplace/redhat-marketplace-gtmnc" Nov 22 06:06:36 crc kubenswrapper[4818]: I1122 06:06:36.754028 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2103acc2-9d59-4f08-97d2-f1cbee955f68-utilities\") pod \"redhat-marketplace-gtmnc\" (UID: \"2103acc2-9d59-4f08-97d2-f1cbee955f68\") " pod="openshift-marketplace/redhat-marketplace-gtmnc" Nov 22 06:06:36 crc kubenswrapper[4818]: I1122 06:06:36.754309 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdz6p\" (UniqueName: \"kubernetes.io/projected/2103acc2-9d59-4f08-97d2-f1cbee955f68-kube-api-access-kdz6p\") pod \"redhat-marketplace-gtmnc\" (UID: \"2103acc2-9d59-4f08-97d2-f1cbee955f68\") " pod="openshift-marketplace/redhat-marketplace-gtmnc" Nov 22 06:06:36 crc kubenswrapper[4818]: I1122 06:06:36.856572 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2103acc2-9d59-4f08-97d2-f1cbee955f68-utilities\") pod \"redhat-marketplace-gtmnc\" (UID: \"2103acc2-9d59-4f08-97d2-f1cbee955f68\") " pod="openshift-marketplace/redhat-marketplace-gtmnc" Nov 22 06:06:36 crc kubenswrapper[4818]: I1122 06:06:36.856770 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdz6p\" (UniqueName: \"kubernetes.io/projected/2103acc2-9d59-4f08-97d2-f1cbee955f68-kube-api-access-kdz6p\") pod \"redhat-marketplace-gtmnc\" (UID: \"2103acc2-9d59-4f08-97d2-f1cbee955f68\") " pod="openshift-marketplace/redhat-marketplace-gtmnc" Nov 22 06:06:36 crc kubenswrapper[4818]: I1122 06:06:36.856932 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2103acc2-9d59-4f08-97d2-f1cbee955f68-catalog-content\") pod \"redhat-marketplace-gtmnc\" (UID: \"2103acc2-9d59-4f08-97d2-f1cbee955f68\") " pod="openshift-marketplace/redhat-marketplace-gtmnc" Nov 22 06:06:36 crc kubenswrapper[4818]: I1122 06:06:36.857292 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2103acc2-9d59-4f08-97d2-f1cbee955f68-utilities\") pod \"redhat-marketplace-gtmnc\" (UID: \"2103acc2-9d59-4f08-97d2-f1cbee955f68\") " pod="openshift-marketplace/redhat-marketplace-gtmnc" Nov 22 06:06:36 crc kubenswrapper[4818]: I1122 06:06:36.857439 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2103acc2-9d59-4f08-97d2-f1cbee955f68-catalog-content\") pod \"redhat-marketplace-gtmnc\" (UID: \"2103acc2-9d59-4f08-97d2-f1cbee955f68\") " pod="openshift-marketplace/redhat-marketplace-gtmnc" Nov 22 06:06:36 crc kubenswrapper[4818]: I1122 06:06:36.889198 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdz6p\" (UniqueName: \"kubernetes.io/projected/2103acc2-9d59-4f08-97d2-f1cbee955f68-kube-api-access-kdz6p\") pod \"redhat-marketplace-gtmnc\" (UID: \"2103acc2-9d59-4f08-97d2-f1cbee955f68\") " pod="openshift-marketplace/redhat-marketplace-gtmnc" Nov 22 06:06:36 crc kubenswrapper[4818]: I1122 06:06:36.998235 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gtmnc" Nov 22 06:06:37 crc kubenswrapper[4818]: W1122 06:06:37.581470 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2103acc2_9d59_4f08_97d2_f1cbee955f68.slice/crio-f317779bbb15a7254cd0c6ee9956274e943f9ed5e298668d555bc6ae47074040 WatchSource:0}: Error finding container f317779bbb15a7254cd0c6ee9956274e943f9ed5e298668d555bc6ae47074040: Status 404 returned error can't find the container with id f317779bbb15a7254cd0c6ee9956274e943f9ed5e298668d555bc6ae47074040 Nov 22 06:06:37 crc kubenswrapper[4818]: I1122 06:06:37.584282 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gtmnc"] Nov 22 06:06:38 crc kubenswrapper[4818]: I1122 06:06:38.319513 4818 generic.go:334] "Generic (PLEG): container finished" podID="2103acc2-9d59-4f08-97d2-f1cbee955f68" containerID="526caed4d0465cc984956417ee541f6488412681645942f7824e14156cb9d1e1" exitCode=0 Nov 22 06:06:38 crc kubenswrapper[4818]: I1122 06:06:38.333478 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtmnc" event={"ID":"2103acc2-9d59-4f08-97d2-f1cbee955f68","Type":"ContainerDied","Data":"526caed4d0465cc984956417ee541f6488412681645942f7824e14156cb9d1e1"} Nov 22 06:06:38 crc kubenswrapper[4818]: I1122 06:06:38.333529 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtmnc" event={"ID":"2103acc2-9d59-4f08-97d2-f1cbee955f68","Type":"ContainerStarted","Data":"f317779bbb15a7254cd0c6ee9956274e943f9ed5e298668d555bc6ae47074040"} Nov 22 06:06:38 crc kubenswrapper[4818]: I1122 06:06:38.370065 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 06:06:39 crc kubenswrapper[4818]: I1122 06:06:39.332001 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtmnc" event={"ID":"2103acc2-9d59-4f08-97d2-f1cbee955f68","Type":"ContainerStarted","Data":"c834ae9a937ea3f7d5da2a69f8f16dbfa83d3b9bdadca123b169cf38cd8feafc"} Nov 22 06:06:40 crc kubenswrapper[4818]: I1122 06:06:40.358377 4818 generic.go:334] "Generic (PLEG): container finished" podID="2103acc2-9d59-4f08-97d2-f1cbee955f68" containerID="c834ae9a937ea3f7d5da2a69f8f16dbfa83d3b9bdadca123b169cf38cd8feafc" exitCode=0 Nov 22 06:06:40 crc kubenswrapper[4818]: I1122 06:06:40.358669 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtmnc" event={"ID":"2103acc2-9d59-4f08-97d2-f1cbee955f68","Type":"ContainerDied","Data":"c834ae9a937ea3f7d5da2a69f8f16dbfa83d3b9bdadca123b169cf38cd8feafc"} Nov 22 06:06:41 crc kubenswrapper[4818]: I1122 06:06:41.368963 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtmnc" event={"ID":"2103acc2-9d59-4f08-97d2-f1cbee955f68","Type":"ContainerStarted","Data":"ff102e5cad7ca3c2d99023dbd51571532eb9072695c689a4671a0a0c8af544a9"} Nov 22 06:06:41 crc kubenswrapper[4818]: I1122 06:06:41.393210 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gtmnc" podStartSLOduration=2.995186564 podStartE2EDuration="5.393192544s" podCreationTimestamp="2025-11-22 06:06:36 +0000 UTC" firstStartedPulling="2025-11-22 06:06:38.369354688 +0000 UTC m=+4750.943771225" lastFinishedPulling="2025-11-22 06:06:40.767360638 +0000 UTC m=+4753.341777205" observedRunningTime="2025-11-22 06:06:41.386876415 +0000 UTC m=+4753.961292952" watchObservedRunningTime="2025-11-22 06:06:41.393192544 +0000 UTC m=+4753.967609071" Nov 22 06:06:45 crc kubenswrapper[4818]: I1122 06:06:45.277931 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-blnwt"] Nov 22 06:06:45 crc kubenswrapper[4818]: I1122 06:06:45.280888 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-blnwt" Nov 22 06:06:45 crc kubenswrapper[4818]: I1122 06:06:45.300650 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-blnwt"] Nov 22 06:06:45 crc kubenswrapper[4818]: I1122 06:06:45.374167 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d219b0-124c-4cb5-809e-4685b4c33ec1-utilities\") pod \"certified-operators-blnwt\" (UID: \"b3d219b0-124c-4cb5-809e-4685b4c33ec1\") " pod="openshift-marketplace/certified-operators-blnwt" Nov 22 06:06:45 crc kubenswrapper[4818]: I1122 06:06:45.374240 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d219b0-124c-4cb5-809e-4685b4c33ec1-catalog-content\") pod \"certified-operators-blnwt\" (UID: \"b3d219b0-124c-4cb5-809e-4685b4c33ec1\") " pod="openshift-marketplace/certified-operators-blnwt" Nov 22 06:06:45 crc kubenswrapper[4818]: I1122 06:06:45.374419 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzjn5\" (UniqueName: \"kubernetes.io/projected/b3d219b0-124c-4cb5-809e-4685b4c33ec1-kube-api-access-qzjn5\") pod \"certified-operators-blnwt\" (UID: \"b3d219b0-124c-4cb5-809e-4685b4c33ec1\") " pod="openshift-marketplace/certified-operators-blnwt" Nov 22 06:06:45 crc kubenswrapper[4818]: I1122 06:06:45.477119 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d219b0-124c-4cb5-809e-4685b4c33ec1-utilities\") pod \"certified-operators-blnwt\" (UID: \"b3d219b0-124c-4cb5-809e-4685b4c33ec1\") " pod="openshift-marketplace/certified-operators-blnwt" Nov 22 06:06:45 crc kubenswrapper[4818]: I1122 06:06:45.477230 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d219b0-124c-4cb5-809e-4685b4c33ec1-catalog-content\") pod \"certified-operators-blnwt\" (UID: \"b3d219b0-124c-4cb5-809e-4685b4c33ec1\") " pod="openshift-marketplace/certified-operators-blnwt" Nov 22 06:06:45 crc kubenswrapper[4818]: I1122 06:06:45.477326 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzjn5\" (UniqueName: \"kubernetes.io/projected/b3d219b0-124c-4cb5-809e-4685b4c33ec1-kube-api-access-qzjn5\") pod \"certified-operators-blnwt\" (UID: \"b3d219b0-124c-4cb5-809e-4685b4c33ec1\") " pod="openshift-marketplace/certified-operators-blnwt" Nov 22 06:06:45 crc kubenswrapper[4818]: I1122 06:06:45.477925 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d219b0-124c-4cb5-809e-4685b4c33ec1-catalog-content\") pod \"certified-operators-blnwt\" (UID: \"b3d219b0-124c-4cb5-809e-4685b4c33ec1\") " pod="openshift-marketplace/certified-operators-blnwt" Nov 22 06:06:45 crc kubenswrapper[4818]: I1122 06:06:45.478079 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d219b0-124c-4cb5-809e-4685b4c33ec1-utilities\") pod \"certified-operators-blnwt\" (UID: \"b3d219b0-124c-4cb5-809e-4685b4c33ec1\") " pod="openshift-marketplace/certified-operators-blnwt" Nov 22 06:06:45 crc kubenswrapper[4818]: I1122 06:06:45.504335 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzjn5\" (UniqueName: \"kubernetes.io/projected/b3d219b0-124c-4cb5-809e-4685b4c33ec1-kube-api-access-qzjn5\") pod \"certified-operators-blnwt\" (UID: \"b3d219b0-124c-4cb5-809e-4685b4c33ec1\") " pod="openshift-marketplace/certified-operators-blnwt" Nov 22 06:06:45 crc kubenswrapper[4818]: I1122 06:06:45.662402 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-blnwt" Nov 22 06:06:46 crc kubenswrapper[4818]: I1122 06:06:46.135218 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-blnwt"] Nov 22 06:06:46 crc kubenswrapper[4818]: I1122 06:06:46.416028 4818 generic.go:334] "Generic (PLEG): container finished" podID="b3d219b0-124c-4cb5-809e-4685b4c33ec1" containerID="47887a05f67c98ac245cd874c7a1981e1d50a83c1420221a980eea11f0163bad" exitCode=0 Nov 22 06:06:46 crc kubenswrapper[4818]: I1122 06:06:46.416092 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blnwt" event={"ID":"b3d219b0-124c-4cb5-809e-4685b4c33ec1","Type":"ContainerDied","Data":"47887a05f67c98ac245cd874c7a1981e1d50a83c1420221a980eea11f0163bad"} Nov 22 06:06:46 crc kubenswrapper[4818]: I1122 06:06:46.416324 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blnwt" event={"ID":"b3d219b0-124c-4cb5-809e-4685b4c33ec1","Type":"ContainerStarted","Data":"68908271e151300736f63c8737b6060853307f004ebdd12eda282862d4c9a4c1"} Nov 22 06:06:46 crc kubenswrapper[4818]: I1122 06:06:46.998640 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gtmnc" Nov 22 06:06:46 crc kubenswrapper[4818]: I1122 06:06:46.999889 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gtmnc" Nov 22 06:06:47 crc kubenswrapper[4818]: I1122 06:06:47.052665 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gtmnc" Nov 22 06:06:47 crc kubenswrapper[4818]: I1122 06:06:47.478237 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gtmnc" Nov 22 06:06:48 crc kubenswrapper[4818]: I1122 06:06:48.437012 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blnwt" event={"ID":"b3d219b0-124c-4cb5-809e-4685b4c33ec1","Type":"ContainerStarted","Data":"b0bf1739a03a4af2454296865607a41b0defe3b0ff60b5fef9650b36bf1f3c39"} Nov 22 06:06:48 crc kubenswrapper[4818]: I1122 06:06:48.659096 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gtmnc"] Nov 22 06:06:49 crc kubenswrapper[4818]: I1122 06:06:49.449056 4818 generic.go:334] "Generic (PLEG): container finished" podID="b3d219b0-124c-4cb5-809e-4685b4c33ec1" containerID="b0bf1739a03a4af2454296865607a41b0defe3b0ff60b5fef9650b36bf1f3c39" exitCode=0 Nov 22 06:06:49 crc kubenswrapper[4818]: I1122 06:06:49.449336 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blnwt" event={"ID":"b3d219b0-124c-4cb5-809e-4685b4c33ec1","Type":"ContainerDied","Data":"b0bf1739a03a4af2454296865607a41b0defe3b0ff60b5fef9650b36bf1f3c39"} Nov 22 06:06:50 crc kubenswrapper[4818]: I1122 06:06:50.460456 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gtmnc" podUID="2103acc2-9d59-4f08-97d2-f1cbee955f68" containerName="registry-server" containerID="cri-o://ff102e5cad7ca3c2d99023dbd51571532eb9072695c689a4671a0a0c8af544a9" gracePeriod=2 Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.298777 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gtmnc" Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.440948 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdz6p\" (UniqueName: \"kubernetes.io/projected/2103acc2-9d59-4f08-97d2-f1cbee955f68-kube-api-access-kdz6p\") pod \"2103acc2-9d59-4f08-97d2-f1cbee955f68\" (UID: \"2103acc2-9d59-4f08-97d2-f1cbee955f68\") " Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.441153 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2103acc2-9d59-4f08-97d2-f1cbee955f68-catalog-content\") pod \"2103acc2-9d59-4f08-97d2-f1cbee955f68\" (UID: \"2103acc2-9d59-4f08-97d2-f1cbee955f68\") " Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.441319 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2103acc2-9d59-4f08-97d2-f1cbee955f68-utilities\") pod \"2103acc2-9d59-4f08-97d2-f1cbee955f68\" (UID: \"2103acc2-9d59-4f08-97d2-f1cbee955f68\") " Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.442161 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2103acc2-9d59-4f08-97d2-f1cbee955f68-utilities" (OuterVolumeSpecName: "utilities") pod "2103acc2-9d59-4f08-97d2-f1cbee955f68" (UID: "2103acc2-9d59-4f08-97d2-f1cbee955f68"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.450334 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2103acc2-9d59-4f08-97d2-f1cbee955f68-kube-api-access-kdz6p" (OuterVolumeSpecName: "kube-api-access-kdz6p") pod "2103acc2-9d59-4f08-97d2-f1cbee955f68" (UID: "2103acc2-9d59-4f08-97d2-f1cbee955f68"). InnerVolumeSpecName "kube-api-access-kdz6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.480032 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2103acc2-9d59-4f08-97d2-f1cbee955f68-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2103acc2-9d59-4f08-97d2-f1cbee955f68" (UID: "2103acc2-9d59-4f08-97d2-f1cbee955f68"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.482848 4818 generic.go:334] "Generic (PLEG): container finished" podID="2103acc2-9d59-4f08-97d2-f1cbee955f68" containerID="ff102e5cad7ca3c2d99023dbd51571532eb9072695c689a4671a0a0c8af544a9" exitCode=0 Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.482922 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtmnc" event={"ID":"2103acc2-9d59-4f08-97d2-f1cbee955f68","Type":"ContainerDied","Data":"ff102e5cad7ca3c2d99023dbd51571532eb9072695c689a4671a0a0c8af544a9"} Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.482955 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtmnc" event={"ID":"2103acc2-9d59-4f08-97d2-f1cbee955f68","Type":"ContainerDied","Data":"f317779bbb15a7254cd0c6ee9956274e943f9ed5e298668d555bc6ae47074040"} Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.482975 4818 scope.go:117] "RemoveContainer" containerID="ff102e5cad7ca3c2d99023dbd51571532eb9072695c689a4671a0a0c8af544a9" Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.483117 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gtmnc" Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.488197 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blnwt" event={"ID":"b3d219b0-124c-4cb5-809e-4685b4c33ec1","Type":"ContainerStarted","Data":"781322f34c2b4065e69b3bc56bb3789b5e4f8a4442caaadb389729e83bfc0cec"} Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.541291 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-blnwt" podStartSLOduration=2.940969246 podStartE2EDuration="6.541273689s" podCreationTimestamp="2025-11-22 06:06:45 +0000 UTC" firstStartedPulling="2025-11-22 06:06:46.418154161 +0000 UTC m=+4758.992570688" lastFinishedPulling="2025-11-22 06:06:50.018458604 +0000 UTC m=+4762.592875131" observedRunningTime="2025-11-22 06:06:51.512627673 +0000 UTC m=+4764.087044220" watchObservedRunningTime="2025-11-22 06:06:51.541273689 +0000 UTC m=+4764.115690216" Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.546636 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2103acc2-9d59-4f08-97d2-f1cbee955f68-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.546677 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2103acc2-9d59-4f08-97d2-f1cbee955f68-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.546691 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdz6p\" (UniqueName: \"kubernetes.io/projected/2103acc2-9d59-4f08-97d2-f1cbee955f68-kube-api-access-kdz6p\") on node \"crc\" DevicePath \"\"" Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.547468 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gtmnc"] Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.549779 4818 scope.go:117] "RemoveContainer" containerID="c834ae9a937ea3f7d5da2a69f8f16dbfa83d3b9bdadca123b169cf38cd8feafc" Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.554866 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gtmnc"] Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.570944 4818 scope.go:117] "RemoveContainer" containerID="526caed4d0465cc984956417ee541f6488412681645942f7824e14156cb9d1e1" Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.621359 4818 scope.go:117] "RemoveContainer" containerID="ff102e5cad7ca3c2d99023dbd51571532eb9072695c689a4671a0a0c8af544a9" Nov 22 06:06:51 crc kubenswrapper[4818]: E1122 06:06:51.621690 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff102e5cad7ca3c2d99023dbd51571532eb9072695c689a4671a0a0c8af544a9\": container with ID starting with ff102e5cad7ca3c2d99023dbd51571532eb9072695c689a4671a0a0c8af544a9 not found: ID does not exist" containerID="ff102e5cad7ca3c2d99023dbd51571532eb9072695c689a4671a0a0c8af544a9" Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.621728 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff102e5cad7ca3c2d99023dbd51571532eb9072695c689a4671a0a0c8af544a9"} err="failed to get container status \"ff102e5cad7ca3c2d99023dbd51571532eb9072695c689a4671a0a0c8af544a9\": rpc error: code = NotFound desc = could not find container \"ff102e5cad7ca3c2d99023dbd51571532eb9072695c689a4671a0a0c8af544a9\": container with ID starting with ff102e5cad7ca3c2d99023dbd51571532eb9072695c689a4671a0a0c8af544a9 not found: ID does not exist" Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.621752 4818 scope.go:117] "RemoveContainer" containerID="c834ae9a937ea3f7d5da2a69f8f16dbfa83d3b9bdadca123b169cf38cd8feafc" Nov 22 06:06:51 crc kubenswrapper[4818]: E1122 06:06:51.621987 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c834ae9a937ea3f7d5da2a69f8f16dbfa83d3b9bdadca123b169cf38cd8feafc\": container with ID starting with c834ae9a937ea3f7d5da2a69f8f16dbfa83d3b9bdadca123b169cf38cd8feafc not found: ID does not exist" containerID="c834ae9a937ea3f7d5da2a69f8f16dbfa83d3b9bdadca123b169cf38cd8feafc" Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.622014 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c834ae9a937ea3f7d5da2a69f8f16dbfa83d3b9bdadca123b169cf38cd8feafc"} err="failed to get container status \"c834ae9a937ea3f7d5da2a69f8f16dbfa83d3b9bdadca123b169cf38cd8feafc\": rpc error: code = NotFound desc = could not find container \"c834ae9a937ea3f7d5da2a69f8f16dbfa83d3b9bdadca123b169cf38cd8feafc\": container with ID starting with c834ae9a937ea3f7d5da2a69f8f16dbfa83d3b9bdadca123b169cf38cd8feafc not found: ID does not exist" Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.622028 4818 scope.go:117] "RemoveContainer" containerID="526caed4d0465cc984956417ee541f6488412681645942f7824e14156cb9d1e1" Nov 22 06:06:51 crc kubenswrapper[4818]: E1122 06:06:51.622284 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"526caed4d0465cc984956417ee541f6488412681645942f7824e14156cb9d1e1\": container with ID starting with 526caed4d0465cc984956417ee541f6488412681645942f7824e14156cb9d1e1 not found: ID does not exist" containerID="526caed4d0465cc984956417ee541f6488412681645942f7824e14156cb9d1e1" Nov 22 06:06:51 crc kubenswrapper[4818]: I1122 06:06:51.622308 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"526caed4d0465cc984956417ee541f6488412681645942f7824e14156cb9d1e1"} err="failed to get container status \"526caed4d0465cc984956417ee541f6488412681645942f7824e14156cb9d1e1\": rpc error: code = NotFound desc = could not find container \"526caed4d0465cc984956417ee541f6488412681645942f7824e14156cb9d1e1\": container with ID starting with 526caed4d0465cc984956417ee541f6488412681645942f7824e14156cb9d1e1 not found: ID does not exist" Nov 22 06:06:52 crc kubenswrapper[4818]: I1122 06:06:52.305497 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2103acc2-9d59-4f08-97d2-f1cbee955f68" path="/var/lib/kubelet/pods/2103acc2-9d59-4f08-97d2-f1cbee955f68/volumes" Nov 22 06:06:55 crc kubenswrapper[4818]: I1122 06:06:55.662945 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-blnwt" Nov 22 06:06:55 crc kubenswrapper[4818]: I1122 06:06:55.663339 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-blnwt" Nov 22 06:06:55 crc kubenswrapper[4818]: I1122 06:06:55.741983 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-blnwt" Nov 22 06:06:56 crc kubenswrapper[4818]: I1122 06:06:56.644357 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-blnwt" Nov 22 06:06:57 crc kubenswrapper[4818]: I1122 06:06:57.059508 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-blnwt"] Nov 22 06:06:58 crc kubenswrapper[4818]: I1122 06:06:58.594759 4818 generic.go:334] "Generic (PLEG): container finished" podID="33278cc5-4458-4244-9cf1-fdbcd53bdda4" containerID="e2f9649b112209878db60859c8f18382ae5b9bf91a3d9413b8445b4093d447ad" exitCode=0 Nov 22 06:06:58 crc kubenswrapper[4818]: I1122 06:06:58.594893 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qz69t/must-gather-qq2wp" event={"ID":"33278cc5-4458-4244-9cf1-fdbcd53bdda4","Type":"ContainerDied","Data":"e2f9649b112209878db60859c8f18382ae5b9bf91a3d9413b8445b4093d447ad"} Nov 22 06:06:58 crc kubenswrapper[4818]: I1122 06:06:58.595712 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-blnwt" podUID="b3d219b0-124c-4cb5-809e-4685b4c33ec1" containerName="registry-server" containerID="cri-o://781322f34c2b4065e69b3bc56bb3789b5e4f8a4442caaadb389729e83bfc0cec" gracePeriod=2 Nov 22 06:06:58 crc kubenswrapper[4818]: I1122 06:06:58.596671 4818 scope.go:117] "RemoveContainer" containerID="e2f9649b112209878db60859c8f18382ae5b9bf91a3d9413b8445b4093d447ad" Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.127674 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-blnwt" Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.327035 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d219b0-124c-4cb5-809e-4685b4c33ec1-utilities\") pod \"b3d219b0-124c-4cb5-809e-4685b4c33ec1\" (UID: \"b3d219b0-124c-4cb5-809e-4685b4c33ec1\") " Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.327681 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d219b0-124c-4cb5-809e-4685b4c33ec1-catalog-content\") pod \"b3d219b0-124c-4cb5-809e-4685b4c33ec1\" (UID: \"b3d219b0-124c-4cb5-809e-4685b4c33ec1\") " Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.327784 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzjn5\" (UniqueName: \"kubernetes.io/projected/b3d219b0-124c-4cb5-809e-4685b4c33ec1-kube-api-access-qzjn5\") pod \"b3d219b0-124c-4cb5-809e-4685b4c33ec1\" (UID: \"b3d219b0-124c-4cb5-809e-4685b4c33ec1\") " Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.327976 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3d219b0-124c-4cb5-809e-4685b4c33ec1-utilities" (OuterVolumeSpecName: "utilities") pod "b3d219b0-124c-4cb5-809e-4685b4c33ec1" (UID: "b3d219b0-124c-4cb5-809e-4685b4c33ec1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.329460 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d219b0-124c-4cb5-809e-4685b4c33ec1-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.334316 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3d219b0-124c-4cb5-809e-4685b4c33ec1-kube-api-access-qzjn5" (OuterVolumeSpecName: "kube-api-access-qzjn5") pod "b3d219b0-124c-4cb5-809e-4685b4c33ec1" (UID: "b3d219b0-124c-4cb5-809e-4685b4c33ec1"). InnerVolumeSpecName "kube-api-access-qzjn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.379045 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qz69t_must-gather-qq2wp_33278cc5-4458-4244-9cf1-fdbcd53bdda4/gather/0.log" Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.385026 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3d219b0-124c-4cb5-809e-4685b4c33ec1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3d219b0-124c-4cb5-809e-4685b4c33ec1" (UID: "b3d219b0-124c-4cb5-809e-4685b4c33ec1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.430751 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d219b0-124c-4cb5-809e-4685b4c33ec1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.430804 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzjn5\" (UniqueName: \"kubernetes.io/projected/b3d219b0-124c-4cb5-809e-4685b4c33ec1-kube-api-access-qzjn5\") on node \"crc\" DevicePath \"\"" Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.608982 4818 generic.go:334] "Generic (PLEG): container finished" podID="b3d219b0-124c-4cb5-809e-4685b4c33ec1" containerID="781322f34c2b4065e69b3bc56bb3789b5e4f8a4442caaadb389729e83bfc0cec" exitCode=0 Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.609036 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blnwt" event={"ID":"b3d219b0-124c-4cb5-809e-4685b4c33ec1","Type":"ContainerDied","Data":"781322f34c2b4065e69b3bc56bb3789b5e4f8a4442caaadb389729e83bfc0cec"} Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.609070 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blnwt" event={"ID":"b3d219b0-124c-4cb5-809e-4685b4c33ec1","Type":"ContainerDied","Data":"68908271e151300736f63c8737b6060853307f004ebdd12eda282862d4c9a4c1"} Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.609125 4818 scope.go:117] "RemoveContainer" containerID="781322f34c2b4065e69b3bc56bb3789b5e4f8a4442caaadb389729e83bfc0cec" Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.609148 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-blnwt" Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.640677 4818 scope.go:117] "RemoveContainer" containerID="b0bf1739a03a4af2454296865607a41b0defe3b0ff60b5fef9650b36bf1f3c39" Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.677249 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-blnwt"] Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.685953 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-blnwt"] Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.685957 4818 scope.go:117] "RemoveContainer" containerID="47887a05f67c98ac245cd874c7a1981e1d50a83c1420221a980eea11f0163bad" Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.720101 4818 scope.go:117] "RemoveContainer" containerID="781322f34c2b4065e69b3bc56bb3789b5e4f8a4442caaadb389729e83bfc0cec" Nov 22 06:06:59 crc kubenswrapper[4818]: E1122 06:06:59.723177 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"781322f34c2b4065e69b3bc56bb3789b5e4f8a4442caaadb389729e83bfc0cec\": container with ID starting with 781322f34c2b4065e69b3bc56bb3789b5e4f8a4442caaadb389729e83bfc0cec not found: ID does not exist" containerID="781322f34c2b4065e69b3bc56bb3789b5e4f8a4442caaadb389729e83bfc0cec" Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.723218 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"781322f34c2b4065e69b3bc56bb3789b5e4f8a4442caaadb389729e83bfc0cec"} err="failed to get container status \"781322f34c2b4065e69b3bc56bb3789b5e4f8a4442caaadb389729e83bfc0cec\": rpc error: code = NotFound desc = could not find container \"781322f34c2b4065e69b3bc56bb3789b5e4f8a4442caaadb389729e83bfc0cec\": container with ID starting with 781322f34c2b4065e69b3bc56bb3789b5e4f8a4442caaadb389729e83bfc0cec not found: ID does not exist" Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.723309 4818 scope.go:117] "RemoveContainer" containerID="b0bf1739a03a4af2454296865607a41b0defe3b0ff60b5fef9650b36bf1f3c39" Nov 22 06:06:59 crc kubenswrapper[4818]: E1122 06:06:59.723977 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0bf1739a03a4af2454296865607a41b0defe3b0ff60b5fef9650b36bf1f3c39\": container with ID starting with b0bf1739a03a4af2454296865607a41b0defe3b0ff60b5fef9650b36bf1f3c39 not found: ID does not exist" containerID="b0bf1739a03a4af2454296865607a41b0defe3b0ff60b5fef9650b36bf1f3c39" Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.724044 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0bf1739a03a4af2454296865607a41b0defe3b0ff60b5fef9650b36bf1f3c39"} err="failed to get container status \"b0bf1739a03a4af2454296865607a41b0defe3b0ff60b5fef9650b36bf1f3c39\": rpc error: code = NotFound desc = could not find container \"b0bf1739a03a4af2454296865607a41b0defe3b0ff60b5fef9650b36bf1f3c39\": container with ID starting with b0bf1739a03a4af2454296865607a41b0defe3b0ff60b5fef9650b36bf1f3c39 not found: ID does not exist" Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.724064 4818 scope.go:117] "RemoveContainer" containerID="47887a05f67c98ac245cd874c7a1981e1d50a83c1420221a980eea11f0163bad" Nov 22 06:06:59 crc kubenswrapper[4818]: E1122 06:06:59.729738 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47887a05f67c98ac245cd874c7a1981e1d50a83c1420221a980eea11f0163bad\": container with ID starting with 47887a05f67c98ac245cd874c7a1981e1d50a83c1420221a980eea11f0163bad not found: ID does not exist" containerID="47887a05f67c98ac245cd874c7a1981e1d50a83c1420221a980eea11f0163bad" Nov 22 06:06:59 crc kubenswrapper[4818]: I1122 06:06:59.729770 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47887a05f67c98ac245cd874c7a1981e1d50a83c1420221a980eea11f0163bad"} err="failed to get container status \"47887a05f67c98ac245cd874c7a1981e1d50a83c1420221a980eea11f0163bad\": rpc error: code = NotFound desc = could not find container \"47887a05f67c98ac245cd874c7a1981e1d50a83c1420221a980eea11f0163bad\": container with ID starting with 47887a05f67c98ac245cd874c7a1981e1d50a83c1420221a980eea11f0163bad not found: ID does not exist" Nov 22 06:07:00 crc kubenswrapper[4818]: I1122 06:07:00.305235 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3d219b0-124c-4cb5-809e-4685b4c33ec1" path="/var/lib/kubelet/pods/b3d219b0-124c-4cb5-809e-4685b4c33ec1/volumes" Nov 22 06:07:07 crc kubenswrapper[4818]: I1122 06:07:07.413006 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qz69t/must-gather-qq2wp"] Nov 22 06:07:07 crc kubenswrapper[4818]: I1122 06:07:07.419375 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-qz69t/must-gather-qq2wp" podUID="33278cc5-4458-4244-9cf1-fdbcd53bdda4" containerName="copy" containerID="cri-o://46d616995bf4f5b5ca92e0e6239884e567f70f2f978fb6e2a99606069aa0ba4e" gracePeriod=2 Nov 22 06:07:07 crc kubenswrapper[4818]: I1122 06:07:07.423788 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qz69t/must-gather-qq2wp"] Nov 22 06:07:07 crc kubenswrapper[4818]: I1122 06:07:07.719144 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qz69t_must-gather-qq2wp_33278cc5-4458-4244-9cf1-fdbcd53bdda4/copy/0.log" Nov 22 06:07:07 crc kubenswrapper[4818]: I1122 06:07:07.719621 4818 generic.go:334] "Generic (PLEG): container finished" podID="33278cc5-4458-4244-9cf1-fdbcd53bdda4" containerID="46d616995bf4f5b5ca92e0e6239884e567f70f2f978fb6e2a99606069aa0ba4e" exitCode=143 Nov 22 06:07:07 crc kubenswrapper[4818]: I1122 06:07:07.933469 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qz69t_must-gather-qq2wp_33278cc5-4458-4244-9cf1-fdbcd53bdda4/copy/0.log" Nov 22 06:07:07 crc kubenswrapper[4818]: I1122 06:07:07.934395 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qz69t/must-gather-qq2wp" Nov 22 06:07:08 crc kubenswrapper[4818]: I1122 06:07:08.023657 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kltf\" (UniqueName: \"kubernetes.io/projected/33278cc5-4458-4244-9cf1-fdbcd53bdda4-kube-api-access-2kltf\") pod \"33278cc5-4458-4244-9cf1-fdbcd53bdda4\" (UID: \"33278cc5-4458-4244-9cf1-fdbcd53bdda4\") " Nov 22 06:07:08 crc kubenswrapper[4818]: I1122 06:07:08.023849 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/33278cc5-4458-4244-9cf1-fdbcd53bdda4-must-gather-output\") pod \"33278cc5-4458-4244-9cf1-fdbcd53bdda4\" (UID: \"33278cc5-4458-4244-9cf1-fdbcd53bdda4\") " Nov 22 06:07:08 crc kubenswrapper[4818]: I1122 06:07:08.029817 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33278cc5-4458-4244-9cf1-fdbcd53bdda4-kube-api-access-2kltf" (OuterVolumeSpecName: "kube-api-access-2kltf") pod "33278cc5-4458-4244-9cf1-fdbcd53bdda4" (UID: "33278cc5-4458-4244-9cf1-fdbcd53bdda4"). InnerVolumeSpecName "kube-api-access-2kltf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 06:07:08 crc kubenswrapper[4818]: I1122 06:07:08.124815 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kltf\" (UniqueName: \"kubernetes.io/projected/33278cc5-4458-4244-9cf1-fdbcd53bdda4-kube-api-access-2kltf\") on node \"crc\" DevicePath \"\"" Nov 22 06:07:08 crc kubenswrapper[4818]: I1122 06:07:08.183033 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33278cc5-4458-4244-9cf1-fdbcd53bdda4-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "33278cc5-4458-4244-9cf1-fdbcd53bdda4" (UID: "33278cc5-4458-4244-9cf1-fdbcd53bdda4"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 06:07:08 crc kubenswrapper[4818]: I1122 06:07:08.226212 4818 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/33278cc5-4458-4244-9cf1-fdbcd53bdda4-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 22 06:07:08 crc kubenswrapper[4818]: I1122 06:07:08.303234 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33278cc5-4458-4244-9cf1-fdbcd53bdda4" path="/var/lib/kubelet/pods/33278cc5-4458-4244-9cf1-fdbcd53bdda4/volumes" Nov 22 06:07:08 crc kubenswrapper[4818]: I1122 06:07:08.733651 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qz69t_must-gather-qq2wp_33278cc5-4458-4244-9cf1-fdbcd53bdda4/copy/0.log" Nov 22 06:07:08 crc kubenswrapper[4818]: I1122 06:07:08.735133 4818 scope.go:117] "RemoveContainer" containerID="46d616995bf4f5b5ca92e0e6239884e567f70f2f978fb6e2a99606069aa0ba4e" Nov 22 06:07:08 crc kubenswrapper[4818]: I1122 06:07:08.735210 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qz69t/must-gather-qq2wp" Nov 22 06:07:08 crc kubenswrapper[4818]: I1122 06:07:08.792802 4818 scope.go:117] "RemoveContainer" containerID="e2f9649b112209878db60859c8f18382ae5b9bf91a3d9413b8445b4093d447ad" Nov 22 06:07:44 crc kubenswrapper[4818]: I1122 06:07:44.780383 4818 scope.go:117] "RemoveContainer" containerID="1b2ae5bbe4d125ca94ae39875d74157be068beb854394f99f99a639ff35e19e1" Nov 22 06:08:44 crc kubenswrapper[4818]: I1122 06:08:44.934749 4818 scope.go:117] "RemoveContainer" containerID="a8e19c4e29b6f5e1bc4dd38e8525ebbb156f594ec527d104c6369385724a5926" Nov 22 06:08:51 crc kubenswrapper[4818]: I1122 06:08:51.264480 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 06:08:51 crc kubenswrapper[4818]: I1122 06:08:51.265095 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 06:09:21 crc kubenswrapper[4818]: I1122 06:09:21.266768 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 06:09:21 crc kubenswrapper[4818]: I1122 06:09:21.267427 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 06:09:49 crc kubenswrapper[4818]: I1122 06:09:49.779114 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bksd9/must-gather-nlvnh"] Nov 22 06:09:49 crc kubenswrapper[4818]: E1122 06:09:49.780140 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d219b0-124c-4cb5-809e-4685b4c33ec1" containerName="extract-utilities" Nov 22 06:09:49 crc kubenswrapper[4818]: I1122 06:09:49.780155 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d219b0-124c-4cb5-809e-4685b4c33ec1" containerName="extract-utilities" Nov 22 06:09:49 crc kubenswrapper[4818]: E1122 06:09:49.780170 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2103acc2-9d59-4f08-97d2-f1cbee955f68" containerName="extract-content" Nov 22 06:09:49 crc kubenswrapper[4818]: I1122 06:09:49.780178 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="2103acc2-9d59-4f08-97d2-f1cbee955f68" containerName="extract-content" Nov 22 06:09:49 crc kubenswrapper[4818]: E1122 06:09:49.780192 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2103acc2-9d59-4f08-97d2-f1cbee955f68" containerName="registry-server" Nov 22 06:09:49 crc kubenswrapper[4818]: I1122 06:09:49.780199 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="2103acc2-9d59-4f08-97d2-f1cbee955f68" containerName="registry-server" Nov 22 06:09:49 crc kubenswrapper[4818]: E1122 06:09:49.780212 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33278cc5-4458-4244-9cf1-fdbcd53bdda4" containerName="copy" Nov 22 06:09:49 crc kubenswrapper[4818]: I1122 06:09:49.780219 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="33278cc5-4458-4244-9cf1-fdbcd53bdda4" containerName="copy" Nov 22 06:09:49 crc kubenswrapper[4818]: E1122 06:09:49.780234 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d219b0-124c-4cb5-809e-4685b4c33ec1" containerName="extract-content" Nov 22 06:09:49 crc kubenswrapper[4818]: I1122 06:09:49.780242 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d219b0-124c-4cb5-809e-4685b4c33ec1" containerName="extract-content" Nov 22 06:09:49 crc kubenswrapper[4818]: E1122 06:09:49.780292 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2103acc2-9d59-4f08-97d2-f1cbee955f68" containerName="extract-utilities" Nov 22 06:09:49 crc kubenswrapper[4818]: I1122 06:09:49.780301 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="2103acc2-9d59-4f08-97d2-f1cbee955f68" containerName="extract-utilities" Nov 22 06:09:49 crc kubenswrapper[4818]: E1122 06:09:49.780313 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33278cc5-4458-4244-9cf1-fdbcd53bdda4" containerName="gather" Nov 22 06:09:49 crc kubenswrapper[4818]: I1122 06:09:49.780320 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="33278cc5-4458-4244-9cf1-fdbcd53bdda4" containerName="gather" Nov 22 06:09:49 crc kubenswrapper[4818]: E1122 06:09:49.780333 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d219b0-124c-4cb5-809e-4685b4c33ec1" containerName="registry-server" Nov 22 06:09:49 crc kubenswrapper[4818]: I1122 06:09:49.780340 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d219b0-124c-4cb5-809e-4685b4c33ec1" containerName="registry-server" Nov 22 06:09:49 crc kubenswrapper[4818]: I1122 06:09:49.780542 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="33278cc5-4458-4244-9cf1-fdbcd53bdda4" containerName="copy" Nov 22 06:09:49 crc kubenswrapper[4818]: I1122 06:09:49.780558 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3d219b0-124c-4cb5-809e-4685b4c33ec1" containerName="registry-server" Nov 22 06:09:49 crc kubenswrapper[4818]: I1122 06:09:49.780574 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="33278cc5-4458-4244-9cf1-fdbcd53bdda4" containerName="gather" Nov 22 06:09:49 crc kubenswrapper[4818]: I1122 06:09:49.780597 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="2103acc2-9d59-4f08-97d2-f1cbee955f68" containerName="registry-server" Nov 22 06:09:49 crc kubenswrapper[4818]: I1122 06:09:49.781942 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bksd9/must-gather-nlvnh" Nov 22 06:09:49 crc kubenswrapper[4818]: I1122 06:09:49.784057 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-bksd9"/"default-dockercfg-jr6zx" Nov 22 06:09:49 crc kubenswrapper[4818]: I1122 06:09:49.787518 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-bksd9"/"openshift-service-ca.crt" Nov 22 06:09:49 crc kubenswrapper[4818]: I1122 06:09:49.788925 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-bksd9"/"kube-root-ca.crt" Nov 22 06:09:49 crc kubenswrapper[4818]: I1122 06:09:49.790745 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-bksd9/must-gather-nlvnh"] Nov 22 06:09:49 crc kubenswrapper[4818]: I1122 06:09:49.900228 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj9ct\" (UniqueName: \"kubernetes.io/projected/cde37c03-b657-46b8-9a24-6ed81eac8c50-kube-api-access-lj9ct\") pod \"must-gather-nlvnh\" (UID: \"cde37c03-b657-46b8-9a24-6ed81eac8c50\") " pod="openshift-must-gather-bksd9/must-gather-nlvnh" Nov 22 06:09:49 crc kubenswrapper[4818]: I1122 06:09:49.900312 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cde37c03-b657-46b8-9a24-6ed81eac8c50-must-gather-output\") pod \"must-gather-nlvnh\" (UID: \"cde37c03-b657-46b8-9a24-6ed81eac8c50\") " pod="openshift-must-gather-bksd9/must-gather-nlvnh" Nov 22 06:09:50 crc kubenswrapper[4818]: I1122 06:09:50.001675 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj9ct\" (UniqueName: \"kubernetes.io/projected/cde37c03-b657-46b8-9a24-6ed81eac8c50-kube-api-access-lj9ct\") pod \"must-gather-nlvnh\" (UID: \"cde37c03-b657-46b8-9a24-6ed81eac8c50\") " pod="openshift-must-gather-bksd9/must-gather-nlvnh" Nov 22 06:09:50 crc kubenswrapper[4818]: I1122 06:09:50.001735 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cde37c03-b657-46b8-9a24-6ed81eac8c50-must-gather-output\") pod \"must-gather-nlvnh\" (UID: \"cde37c03-b657-46b8-9a24-6ed81eac8c50\") " pod="openshift-must-gather-bksd9/must-gather-nlvnh" Nov 22 06:09:50 crc kubenswrapper[4818]: I1122 06:09:50.002098 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cde37c03-b657-46b8-9a24-6ed81eac8c50-must-gather-output\") pod \"must-gather-nlvnh\" (UID: \"cde37c03-b657-46b8-9a24-6ed81eac8c50\") " pod="openshift-must-gather-bksd9/must-gather-nlvnh" Nov 22 06:09:50 crc kubenswrapper[4818]: I1122 06:09:50.024091 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj9ct\" (UniqueName: \"kubernetes.io/projected/cde37c03-b657-46b8-9a24-6ed81eac8c50-kube-api-access-lj9ct\") pod \"must-gather-nlvnh\" (UID: \"cde37c03-b657-46b8-9a24-6ed81eac8c50\") " pod="openshift-must-gather-bksd9/must-gather-nlvnh" Nov 22 06:09:50 crc kubenswrapper[4818]: I1122 06:09:50.100045 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bksd9/must-gather-nlvnh" Nov 22 06:09:50 crc kubenswrapper[4818]: I1122 06:09:50.607803 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-bksd9/must-gather-nlvnh"] Nov 22 06:09:50 crc kubenswrapper[4818]: W1122 06:09:50.624325 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcde37c03_b657_46b8_9a24_6ed81eac8c50.slice/crio-7e4f4c45115ab5e47e558f1dd50a7bbb39b18cef4e9baed85f2914b3a1a87c59 WatchSource:0}: Error finding container 7e4f4c45115ab5e47e558f1dd50a7bbb39b18cef4e9baed85f2914b3a1a87c59: Status 404 returned error can't find the container with id 7e4f4c45115ab5e47e558f1dd50a7bbb39b18cef4e9baed85f2914b3a1a87c59 Nov 22 06:09:50 crc kubenswrapper[4818]: I1122 06:09:50.651969 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bksd9/must-gather-nlvnh" event={"ID":"cde37c03-b657-46b8-9a24-6ed81eac8c50","Type":"ContainerStarted","Data":"7e4f4c45115ab5e47e558f1dd50a7bbb39b18cef4e9baed85f2914b3a1a87c59"} Nov 22 06:09:51 crc kubenswrapper[4818]: I1122 06:09:51.972704 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 06:09:51 crc kubenswrapper[4818]: I1122 06:09:51.973068 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 06:09:51 crc kubenswrapper[4818]: I1122 06:09:51.973134 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 06:09:51 crc kubenswrapper[4818]: I1122 06:09:51.980900 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d2100269659d0a3f462da37bed1df37938f3da7494989b3a814f1fdfb9d5a0f9"} pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 06:09:51 crc kubenswrapper[4818]: I1122 06:09:51.981027 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" containerID="cri-o://d2100269659d0a3f462da37bed1df37938f3da7494989b3a814f1fdfb9d5a0f9" gracePeriod=600 Nov 22 06:09:52 crc kubenswrapper[4818]: I1122 06:09:52.071999 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bksd9/must-gather-nlvnh" event={"ID":"cde37c03-b657-46b8-9a24-6ed81eac8c50","Type":"ContainerStarted","Data":"976cfeb045a0f62146acce89fab7e0d55deb6a063823564ae8c401f996757811"} Nov 22 06:09:53 crc kubenswrapper[4818]: I1122 06:09:53.088183 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bksd9/must-gather-nlvnh" event={"ID":"cde37c03-b657-46b8-9a24-6ed81eac8c50","Type":"ContainerStarted","Data":"e56317ded1280e57b749879ff256808bebaa3a20ae0c363fb055b08ce17c79a1"} Nov 22 06:09:53 crc kubenswrapper[4818]: I1122 06:09:53.094508 4818 generic.go:334] "Generic (PLEG): container finished" podID="0971c9ec-29ce-44eb-9875-ad4664143305" containerID="d2100269659d0a3f462da37bed1df37938f3da7494989b3a814f1fdfb9d5a0f9" exitCode=0 Nov 22 06:09:53 crc kubenswrapper[4818]: I1122 06:09:53.094585 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerDied","Data":"d2100269659d0a3f462da37bed1df37938f3da7494989b3a814f1fdfb9d5a0f9"} Nov 22 06:09:53 crc kubenswrapper[4818]: I1122 06:09:53.094634 4818 scope.go:117] "RemoveContainer" containerID="0873ed03814b1a51c7aa344d3f193ef7380552ceaed366ba91d6af50d6ad2861" Nov 22 06:09:53 crc kubenswrapper[4818]: I1122 06:09:53.109501 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-bksd9/must-gather-nlvnh" podStartSLOduration=4.109471824 podStartE2EDuration="4.109471824s" podCreationTimestamp="2025-11-22 06:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 06:09:53.106114374 +0000 UTC m=+4945.680530951" watchObservedRunningTime="2025-11-22 06:09:53.109471824 +0000 UTC m=+4945.683888401" Nov 22 06:09:54 crc kubenswrapper[4818]: I1122 06:09:54.110978 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa"} Nov 22 06:09:56 crc kubenswrapper[4818]: I1122 06:09:56.407825 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bksd9/crc-debug-z2thf"] Nov 22 06:09:56 crc kubenswrapper[4818]: I1122 06:09:56.410233 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bksd9/crc-debug-z2thf" Nov 22 06:09:56 crc kubenswrapper[4818]: I1122 06:09:56.491550 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6blx\" (UniqueName: \"kubernetes.io/projected/b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c-kube-api-access-n6blx\") pod \"crc-debug-z2thf\" (UID: \"b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c\") " pod="openshift-must-gather-bksd9/crc-debug-z2thf" Nov 22 06:09:56 crc kubenswrapper[4818]: I1122 06:09:56.491744 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c-host\") pod \"crc-debug-z2thf\" (UID: \"b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c\") " pod="openshift-must-gather-bksd9/crc-debug-z2thf" Nov 22 06:09:56 crc kubenswrapper[4818]: I1122 06:09:56.593018 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6blx\" (UniqueName: \"kubernetes.io/projected/b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c-kube-api-access-n6blx\") pod \"crc-debug-z2thf\" (UID: \"b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c\") " pod="openshift-must-gather-bksd9/crc-debug-z2thf" Nov 22 06:09:56 crc kubenswrapper[4818]: I1122 06:09:56.593464 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c-host\") pod \"crc-debug-z2thf\" (UID: \"b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c\") " pod="openshift-must-gather-bksd9/crc-debug-z2thf" Nov 22 06:09:56 crc kubenswrapper[4818]: I1122 06:09:56.593603 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c-host\") pod \"crc-debug-z2thf\" (UID: \"b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c\") " pod="openshift-must-gather-bksd9/crc-debug-z2thf" Nov 22 06:09:56 crc kubenswrapper[4818]: I1122 06:09:56.622414 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6blx\" (UniqueName: \"kubernetes.io/projected/b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c-kube-api-access-n6blx\") pod \"crc-debug-z2thf\" (UID: \"b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c\") " pod="openshift-must-gather-bksd9/crc-debug-z2thf" Nov 22 06:09:56 crc kubenswrapper[4818]: I1122 06:09:56.729133 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bksd9/crc-debug-z2thf" Nov 22 06:09:57 crc kubenswrapper[4818]: I1122 06:09:57.137876 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bksd9/crc-debug-z2thf" event={"ID":"b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c","Type":"ContainerStarted","Data":"42abf424655aa3807cdf8f756609f6640346b670aac5ee436abb18403dd5819b"} Nov 22 06:09:58 crc kubenswrapper[4818]: I1122 06:09:58.148057 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bksd9/crc-debug-z2thf" event={"ID":"b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c","Type":"ContainerStarted","Data":"75ab0f23e878da9e10791027425c354bd8679100d4614b7881c89ff31cf69cd5"} Nov 22 06:09:58 crc kubenswrapper[4818]: I1122 06:09:58.167317 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-bksd9/crc-debug-z2thf" podStartSLOduration=2.167300834 podStartE2EDuration="2.167300834s" podCreationTimestamp="2025-11-22 06:09:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 06:09:58.163310287 +0000 UTC m=+4950.737726814" watchObservedRunningTime="2025-11-22 06:09:58.167300834 +0000 UTC m=+4950.741717351" Nov 22 06:10:36 crc kubenswrapper[4818]: I1122 06:10:36.491971 4818 generic.go:334] "Generic (PLEG): container finished" podID="b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c" containerID="75ab0f23e878da9e10791027425c354bd8679100d4614b7881c89ff31cf69cd5" exitCode=0 Nov 22 06:10:36 crc kubenswrapper[4818]: I1122 06:10:36.492057 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bksd9/crc-debug-z2thf" event={"ID":"b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c","Type":"ContainerDied","Data":"75ab0f23e878da9e10791027425c354bd8679100d4614b7881c89ff31cf69cd5"} Nov 22 06:10:37 crc kubenswrapper[4818]: I1122 06:10:37.648217 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bksd9/crc-debug-z2thf" Nov 22 06:10:37 crc kubenswrapper[4818]: I1122 06:10:37.687400 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bksd9/crc-debug-z2thf"] Nov 22 06:10:37 crc kubenswrapper[4818]: I1122 06:10:37.696876 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bksd9/crc-debug-z2thf"] Nov 22 06:10:37 crc kubenswrapper[4818]: I1122 06:10:37.785074 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6blx\" (UniqueName: \"kubernetes.io/projected/b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c-kube-api-access-n6blx\") pod \"b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c\" (UID: \"b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c\") " Nov 22 06:10:37 crc kubenswrapper[4818]: I1122 06:10:37.785137 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c-host\") pod \"b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c\" (UID: \"b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c\") " Nov 22 06:10:37 crc kubenswrapper[4818]: I1122 06:10:37.785163 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c-host" (OuterVolumeSpecName: "host") pod "b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c" (UID: "b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 06:10:37 crc kubenswrapper[4818]: I1122 06:10:37.785635 4818 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c-host\") on node \"crc\" DevicePath \"\"" Nov 22 06:10:37 crc kubenswrapper[4818]: I1122 06:10:37.994822 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c-kube-api-access-n6blx" (OuterVolumeSpecName: "kube-api-access-n6blx") pod "b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c" (UID: "b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c"). InnerVolumeSpecName "kube-api-access-n6blx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 06:10:38 crc kubenswrapper[4818]: I1122 06:10:38.093509 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6blx\" (UniqueName: \"kubernetes.io/projected/b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c-kube-api-access-n6blx\") on node \"crc\" DevicePath \"\"" Nov 22 06:10:38 crc kubenswrapper[4818]: I1122 06:10:38.304044 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c" path="/var/lib/kubelet/pods/b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c/volumes" Nov 22 06:10:38 crc kubenswrapper[4818]: I1122 06:10:38.511689 4818 scope.go:117] "RemoveContainer" containerID="75ab0f23e878da9e10791027425c354bd8679100d4614b7881c89ff31cf69cd5" Nov 22 06:10:38 crc kubenswrapper[4818]: I1122 06:10:38.511704 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bksd9/crc-debug-z2thf" Nov 22 06:10:38 crc kubenswrapper[4818]: I1122 06:10:38.902071 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bksd9/crc-debug-4xwt7"] Nov 22 06:10:38 crc kubenswrapper[4818]: E1122 06:10:38.902583 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c" containerName="container-00" Nov 22 06:10:38 crc kubenswrapper[4818]: I1122 06:10:38.902600 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c" containerName="container-00" Nov 22 06:10:38 crc kubenswrapper[4818]: I1122 06:10:38.902866 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9bc2aa3-f653-4a3b-9c2f-9f2b675b0c9c" containerName="container-00" Nov 22 06:10:38 crc kubenswrapper[4818]: I1122 06:10:38.903628 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bksd9/crc-debug-4xwt7" Nov 22 06:10:39 crc kubenswrapper[4818]: I1122 06:10:39.009968 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1179caa7-1733-4e27-86e6-34bddf3d3375-host\") pod \"crc-debug-4xwt7\" (UID: \"1179caa7-1733-4e27-86e6-34bddf3d3375\") " pod="openshift-must-gather-bksd9/crc-debug-4xwt7" Nov 22 06:10:39 crc kubenswrapper[4818]: I1122 06:10:39.010054 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2ftg\" (UniqueName: \"kubernetes.io/projected/1179caa7-1733-4e27-86e6-34bddf3d3375-kube-api-access-z2ftg\") pod \"crc-debug-4xwt7\" (UID: \"1179caa7-1733-4e27-86e6-34bddf3d3375\") " pod="openshift-must-gather-bksd9/crc-debug-4xwt7" Nov 22 06:10:39 crc kubenswrapper[4818]: I1122 06:10:39.112092 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1179caa7-1733-4e27-86e6-34bddf3d3375-host\") pod \"crc-debug-4xwt7\" (UID: \"1179caa7-1733-4e27-86e6-34bddf3d3375\") " pod="openshift-must-gather-bksd9/crc-debug-4xwt7" Nov 22 06:10:39 crc kubenswrapper[4818]: I1122 06:10:39.112142 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2ftg\" (UniqueName: \"kubernetes.io/projected/1179caa7-1733-4e27-86e6-34bddf3d3375-kube-api-access-z2ftg\") pod \"crc-debug-4xwt7\" (UID: \"1179caa7-1733-4e27-86e6-34bddf3d3375\") " pod="openshift-must-gather-bksd9/crc-debug-4xwt7" Nov 22 06:10:39 crc kubenswrapper[4818]: I1122 06:10:39.112296 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1179caa7-1733-4e27-86e6-34bddf3d3375-host\") pod \"crc-debug-4xwt7\" (UID: \"1179caa7-1733-4e27-86e6-34bddf3d3375\") " pod="openshift-must-gather-bksd9/crc-debug-4xwt7" Nov 22 06:10:39 crc kubenswrapper[4818]: I1122 06:10:39.132031 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2ftg\" (UniqueName: \"kubernetes.io/projected/1179caa7-1733-4e27-86e6-34bddf3d3375-kube-api-access-z2ftg\") pod \"crc-debug-4xwt7\" (UID: \"1179caa7-1733-4e27-86e6-34bddf3d3375\") " pod="openshift-must-gather-bksd9/crc-debug-4xwt7" Nov 22 06:10:39 crc kubenswrapper[4818]: I1122 06:10:39.223035 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bksd9/crc-debug-4xwt7" Nov 22 06:10:39 crc kubenswrapper[4818]: I1122 06:10:39.522512 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bksd9/crc-debug-4xwt7" event={"ID":"1179caa7-1733-4e27-86e6-34bddf3d3375","Type":"ContainerStarted","Data":"066565a765efb2837dc925c23292bb06f36dda8c572718dd171f28df7c8dae76"} Nov 22 06:10:40 crc kubenswrapper[4818]: I1122 06:10:40.533627 4818 generic.go:334] "Generic (PLEG): container finished" podID="1179caa7-1733-4e27-86e6-34bddf3d3375" containerID="aee93133c4db90d2f89f1213881b6fec798bfff424107a513a4b36de4266eaf1" exitCode=0 Nov 22 06:10:40 crc kubenswrapper[4818]: I1122 06:10:40.533722 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bksd9/crc-debug-4xwt7" event={"ID":"1179caa7-1733-4e27-86e6-34bddf3d3375","Type":"ContainerDied","Data":"aee93133c4db90d2f89f1213881b6fec798bfff424107a513a4b36de4266eaf1"} Nov 22 06:10:41 crc kubenswrapper[4818]: I1122 06:10:41.665914 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bksd9/crc-debug-4xwt7" Nov 22 06:10:41 crc kubenswrapper[4818]: I1122 06:10:41.764307 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2ftg\" (UniqueName: \"kubernetes.io/projected/1179caa7-1733-4e27-86e6-34bddf3d3375-kube-api-access-z2ftg\") pod \"1179caa7-1733-4e27-86e6-34bddf3d3375\" (UID: \"1179caa7-1733-4e27-86e6-34bddf3d3375\") " Nov 22 06:10:41 crc kubenswrapper[4818]: I1122 06:10:41.764493 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1179caa7-1733-4e27-86e6-34bddf3d3375-host\") pod \"1179caa7-1733-4e27-86e6-34bddf3d3375\" (UID: \"1179caa7-1733-4e27-86e6-34bddf3d3375\") " Nov 22 06:10:41 crc kubenswrapper[4818]: I1122 06:10:41.764607 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1179caa7-1733-4e27-86e6-34bddf3d3375-host" (OuterVolumeSpecName: "host") pod "1179caa7-1733-4e27-86e6-34bddf3d3375" (UID: "1179caa7-1733-4e27-86e6-34bddf3d3375"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 06:10:41 crc kubenswrapper[4818]: I1122 06:10:41.766770 4818 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1179caa7-1733-4e27-86e6-34bddf3d3375-host\") on node \"crc\" DevicePath \"\"" Nov 22 06:10:41 crc kubenswrapper[4818]: I1122 06:10:41.772760 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1179caa7-1733-4e27-86e6-34bddf3d3375-kube-api-access-z2ftg" (OuterVolumeSpecName: "kube-api-access-z2ftg") pod "1179caa7-1733-4e27-86e6-34bddf3d3375" (UID: "1179caa7-1733-4e27-86e6-34bddf3d3375"). InnerVolumeSpecName "kube-api-access-z2ftg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 06:10:41 crc kubenswrapper[4818]: I1122 06:10:41.870406 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2ftg\" (UniqueName: \"kubernetes.io/projected/1179caa7-1733-4e27-86e6-34bddf3d3375-kube-api-access-z2ftg\") on node \"crc\" DevicePath \"\"" Nov 22 06:10:42 crc kubenswrapper[4818]: I1122 06:10:42.209846 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bksd9/crc-debug-4xwt7"] Nov 22 06:10:42 crc kubenswrapper[4818]: I1122 06:10:42.218727 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bksd9/crc-debug-4xwt7"] Nov 22 06:10:42 crc kubenswrapper[4818]: I1122 06:10:42.300638 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1179caa7-1733-4e27-86e6-34bddf3d3375" path="/var/lib/kubelet/pods/1179caa7-1733-4e27-86e6-34bddf3d3375/volumes" Nov 22 06:10:42 crc kubenswrapper[4818]: I1122 06:10:42.578185 4818 scope.go:117] "RemoveContainer" containerID="aee93133c4db90d2f89f1213881b6fec798bfff424107a513a4b36de4266eaf1" Nov 22 06:10:42 crc kubenswrapper[4818]: I1122 06:10:42.578380 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bksd9/crc-debug-4xwt7" Nov 22 06:10:43 crc kubenswrapper[4818]: I1122 06:10:43.424940 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bksd9/crc-debug-gh5zj"] Nov 22 06:10:43 crc kubenswrapper[4818]: E1122 06:10:43.425672 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1179caa7-1733-4e27-86e6-34bddf3d3375" containerName="container-00" Nov 22 06:10:43 crc kubenswrapper[4818]: I1122 06:10:43.425689 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="1179caa7-1733-4e27-86e6-34bddf3d3375" containerName="container-00" Nov 22 06:10:43 crc kubenswrapper[4818]: I1122 06:10:43.425992 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="1179caa7-1733-4e27-86e6-34bddf3d3375" containerName="container-00" Nov 22 06:10:43 crc kubenswrapper[4818]: I1122 06:10:43.426801 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bksd9/crc-debug-gh5zj" Nov 22 06:10:43 crc kubenswrapper[4818]: I1122 06:10:43.507701 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/39def298-f711-4bcb-8fee-e95d7c37c606-host\") pod \"crc-debug-gh5zj\" (UID: \"39def298-f711-4bcb-8fee-e95d7c37c606\") " pod="openshift-must-gather-bksd9/crc-debug-gh5zj" Nov 22 06:10:43 crc kubenswrapper[4818]: I1122 06:10:43.507873 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgpjm\" (UniqueName: \"kubernetes.io/projected/39def298-f711-4bcb-8fee-e95d7c37c606-kube-api-access-kgpjm\") pod \"crc-debug-gh5zj\" (UID: \"39def298-f711-4bcb-8fee-e95d7c37c606\") " pod="openshift-must-gather-bksd9/crc-debug-gh5zj" Nov 22 06:10:43 crc kubenswrapper[4818]: I1122 06:10:43.609604 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgpjm\" (UniqueName: \"kubernetes.io/projected/39def298-f711-4bcb-8fee-e95d7c37c606-kube-api-access-kgpjm\") pod \"crc-debug-gh5zj\" (UID: \"39def298-f711-4bcb-8fee-e95d7c37c606\") " pod="openshift-must-gather-bksd9/crc-debug-gh5zj" Nov 22 06:10:43 crc kubenswrapper[4818]: I1122 06:10:43.609821 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/39def298-f711-4bcb-8fee-e95d7c37c606-host\") pod \"crc-debug-gh5zj\" (UID: \"39def298-f711-4bcb-8fee-e95d7c37c606\") " pod="openshift-must-gather-bksd9/crc-debug-gh5zj" Nov 22 06:10:43 crc kubenswrapper[4818]: I1122 06:10:43.609946 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/39def298-f711-4bcb-8fee-e95d7c37c606-host\") pod \"crc-debug-gh5zj\" (UID: \"39def298-f711-4bcb-8fee-e95d7c37c606\") " pod="openshift-must-gather-bksd9/crc-debug-gh5zj" Nov 22 06:10:43 crc kubenswrapper[4818]: I1122 06:10:43.645319 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgpjm\" (UniqueName: \"kubernetes.io/projected/39def298-f711-4bcb-8fee-e95d7c37c606-kube-api-access-kgpjm\") pod \"crc-debug-gh5zj\" (UID: \"39def298-f711-4bcb-8fee-e95d7c37c606\") " pod="openshift-must-gather-bksd9/crc-debug-gh5zj" Nov 22 06:10:43 crc kubenswrapper[4818]: I1122 06:10:43.747805 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bksd9/crc-debug-gh5zj" Nov 22 06:10:43 crc kubenswrapper[4818]: W1122 06:10:43.790935 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39def298_f711_4bcb_8fee_e95d7c37c606.slice/crio-a4451fbcae7394be25886856c9305046041111e0adbc22f8fcbd1d73c4ed063c WatchSource:0}: Error finding container a4451fbcae7394be25886856c9305046041111e0adbc22f8fcbd1d73c4ed063c: Status 404 returned error can't find the container with id a4451fbcae7394be25886856c9305046041111e0adbc22f8fcbd1d73c4ed063c Nov 22 06:10:44 crc kubenswrapper[4818]: I1122 06:10:44.598746 4818 generic.go:334] "Generic (PLEG): container finished" podID="39def298-f711-4bcb-8fee-e95d7c37c606" containerID="02c3e02a2c4ea524c736f6423b063c7d5c0e839bc4400e52d99651a5158a0c42" exitCode=0 Nov 22 06:10:44 crc kubenswrapper[4818]: I1122 06:10:44.598817 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bksd9/crc-debug-gh5zj" event={"ID":"39def298-f711-4bcb-8fee-e95d7c37c606","Type":"ContainerDied","Data":"02c3e02a2c4ea524c736f6423b063c7d5c0e839bc4400e52d99651a5158a0c42"} Nov 22 06:10:44 crc kubenswrapper[4818]: I1122 06:10:44.599342 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bksd9/crc-debug-gh5zj" event={"ID":"39def298-f711-4bcb-8fee-e95d7c37c606","Type":"ContainerStarted","Data":"a4451fbcae7394be25886856c9305046041111e0adbc22f8fcbd1d73c4ed063c"} Nov 22 06:10:44 crc kubenswrapper[4818]: I1122 06:10:44.634030 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bksd9/crc-debug-gh5zj"] Nov 22 06:10:44 crc kubenswrapper[4818]: I1122 06:10:44.643877 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bksd9/crc-debug-gh5zj"] Nov 22 06:10:45 crc kubenswrapper[4818]: I1122 06:10:45.744233 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bksd9/crc-debug-gh5zj" Nov 22 06:10:45 crc kubenswrapper[4818]: I1122 06:10:45.854042 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgpjm\" (UniqueName: \"kubernetes.io/projected/39def298-f711-4bcb-8fee-e95d7c37c606-kube-api-access-kgpjm\") pod \"39def298-f711-4bcb-8fee-e95d7c37c606\" (UID: \"39def298-f711-4bcb-8fee-e95d7c37c606\") " Nov 22 06:10:45 crc kubenswrapper[4818]: I1122 06:10:45.854196 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/39def298-f711-4bcb-8fee-e95d7c37c606-host\") pod \"39def298-f711-4bcb-8fee-e95d7c37c606\" (UID: \"39def298-f711-4bcb-8fee-e95d7c37c606\") " Nov 22 06:10:45 crc kubenswrapper[4818]: I1122 06:10:45.854927 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/39def298-f711-4bcb-8fee-e95d7c37c606-host" (OuterVolumeSpecName: "host") pod "39def298-f711-4bcb-8fee-e95d7c37c606" (UID: "39def298-f711-4bcb-8fee-e95d7c37c606"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 06:10:45 crc kubenswrapper[4818]: I1122 06:10:45.872884 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39def298-f711-4bcb-8fee-e95d7c37c606-kube-api-access-kgpjm" (OuterVolumeSpecName: "kube-api-access-kgpjm") pod "39def298-f711-4bcb-8fee-e95d7c37c606" (UID: "39def298-f711-4bcb-8fee-e95d7c37c606"). InnerVolumeSpecName "kube-api-access-kgpjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 06:10:45 crc kubenswrapper[4818]: I1122 06:10:45.956403 4818 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/39def298-f711-4bcb-8fee-e95d7c37c606-host\") on node \"crc\" DevicePath \"\"" Nov 22 06:10:45 crc kubenswrapper[4818]: I1122 06:10:45.956439 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgpjm\" (UniqueName: \"kubernetes.io/projected/39def298-f711-4bcb-8fee-e95d7c37c606-kube-api-access-kgpjm\") on node \"crc\" DevicePath \"\"" Nov 22 06:10:46 crc kubenswrapper[4818]: I1122 06:10:46.301377 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39def298-f711-4bcb-8fee-e95d7c37c606" path="/var/lib/kubelet/pods/39def298-f711-4bcb-8fee-e95d7c37c606/volumes" Nov 22 06:10:46 crc kubenswrapper[4818]: I1122 06:10:46.631972 4818 scope.go:117] "RemoveContainer" containerID="02c3e02a2c4ea524c736f6423b063c7d5c0e839bc4400e52d99651a5158a0c42" Nov 22 06:10:46 crc kubenswrapper[4818]: I1122 06:10:46.632033 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bksd9/crc-debug-gh5zj" Nov 22 06:11:35 crc kubenswrapper[4818]: I1122 06:11:35.575069 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6dfb48dcf4-6ph2v_1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d/barbican-api/0.log" Nov 22 06:11:35 crc kubenswrapper[4818]: I1122 06:11:35.579999 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6dfb48dcf4-6ph2v_1e1ed319-7d7b-4a93-925c-c7bf4bbfaa1d/barbican-api-log/0.log" Nov 22 06:11:35 crc kubenswrapper[4818]: I1122 06:11:35.750013 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5c65499f7d-s6w7z_a5925d65-97cd-493a-8114-6ed07de4386d/barbican-keystone-listener/0.log" Nov 22 06:11:35 crc kubenswrapper[4818]: I1122 06:11:35.879176 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-fb9947b9-nbdwz_3958bab3-ce66-4d2d-8a76-75e1b7593199/barbican-worker/0.log" Nov 22 06:11:36 crc kubenswrapper[4818]: I1122 06:11:36.008686 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-fb9947b9-nbdwz_3958bab3-ce66-4d2d-8a76-75e1b7593199/barbican-worker-log/0.log" Nov 22 06:11:36 crc kubenswrapper[4818]: I1122 06:11:36.042793 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5c65499f7d-s6w7z_a5925d65-97cd-493a-8114-6ed07de4386d/barbican-keystone-listener-log/0.log" Nov 22 06:11:36 crc kubenswrapper[4818]: I1122 06:11:36.155652 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-d8d8g_597bc073-9fa3-485c-b23d-8f7b5ea21da9/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:11:36 crc kubenswrapper[4818]: I1122 06:11:36.227908 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4ea3ac76-3418-4b1c-8800-d57f554e50e3/ceilometer-central-agent/0.log" Nov 22 06:11:36 crc kubenswrapper[4818]: I1122 06:11:36.293147 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4ea3ac76-3418-4b1c-8800-d57f554e50e3/ceilometer-notification-agent/0.log" Nov 22 06:11:36 crc kubenswrapper[4818]: I1122 06:11:36.370914 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4ea3ac76-3418-4b1c-8800-d57f554e50e3/proxy-httpd/0.log" Nov 22 06:11:36 crc kubenswrapper[4818]: I1122 06:11:36.390546 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4ea3ac76-3418-4b1c-8800-d57f554e50e3/sg-core/0.log" Nov 22 06:11:36 crc kubenswrapper[4818]: I1122 06:11:36.481769 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-jkmn2_1488ee6c-884b-4a61-93c9-ae582c9f4778/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:11:36 crc kubenswrapper[4818]: I1122 06:11:36.562446 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lvvb8_5fe8b800-5e58-48ae-8da3-eae014028909/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:11:36 crc kubenswrapper[4818]: I1122 06:11:36.739232 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9a4f90e5-006b-43db-b910-429b496a7c65/cinder-api-log/0.log" Nov 22 06:11:36 crc kubenswrapper[4818]: I1122 06:11:36.751858 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9a4f90e5-006b-43db-b910-429b496a7c65/cinder-api/0.log" Nov 22 06:11:36 crc kubenswrapper[4818]: I1122 06:11:36.912293 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_0c1e7078-787f-4725-80b1-45f7047e108f/probe/0.log" Nov 22 06:11:37 crc kubenswrapper[4818]: I1122 06:11:37.119356 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_0fce1ca4-c9be-4e2a-947a-3cc652edeb50/cinder-scheduler/0.log" Nov 22 06:11:37 crc kubenswrapper[4818]: I1122 06:11:37.161602 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_0fce1ca4-c9be-4e2a-947a-3cc652edeb50/probe/0.log" Nov 22 06:11:37 crc kubenswrapper[4818]: I1122 06:11:37.183321 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_0c1e7078-787f-4725-80b1-45f7047e108f/cinder-backup/0.log" Nov 22 06:11:37 crc kubenswrapper[4818]: I1122 06:11:37.374026 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_3dfef980-4eab-42b9-821e-233a264843cf/cinder-volume/0.log" Nov 22 06:11:37 crc kubenswrapper[4818]: I1122 06:11:37.405875 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-2dxj6_4e6cd425-5f7c-4f17-864b-2e3d62851c58/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:11:37 crc kubenswrapper[4818]: I1122 06:11:37.416168 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_3dfef980-4eab-42b9-821e-233a264843cf/probe/0.log" Nov 22 06:11:37 crc kubenswrapper[4818]: I1122 06:11:37.619532 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-p5t9z_d3f48375-4d0b-4876-b4e8-64a9fa970b63/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:11:37 crc kubenswrapper[4818]: I1122 06:11:37.643840 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-zl4gj_cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b/init/0.log" Nov 22 06:11:37 crc kubenswrapper[4818]: I1122 06:11:37.842213 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-zl4gj_cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b/init/0.log" Nov 22 06:11:37 crc kubenswrapper[4818]: I1122 06:11:37.890829 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_6652494e-6fe5-4c2b-b265-bb2e5567265e/glance-httpd/0.log" Nov 22 06:11:37 crc kubenswrapper[4818]: I1122 06:11:37.894107 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-zl4gj_cdc470d7-9aa0-4a4e-ac5b-c2f97d8f130b/dnsmasq-dns/0.log" Nov 22 06:11:38 crc kubenswrapper[4818]: I1122 06:11:38.024749 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_6652494e-6fe5-4c2b-b265-bb2e5567265e/glance-log/0.log" Nov 22 06:11:38 crc kubenswrapper[4818]: I1122 06:11:38.069173 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f87508a5-9796-46dd-b9a4-29815fd608ad/glance-httpd/0.log" Nov 22 06:11:38 crc kubenswrapper[4818]: I1122 06:11:38.134465 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f87508a5-9796-46dd-b9a4-29815fd608ad/glance-log/0.log" Nov 22 06:11:38 crc kubenswrapper[4818]: I1122 06:11:38.354997 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-86757bfcb-s9bbk_34f2acb8-906b-4520-86c2-bb603a964489/horizon/0.log" Nov 22 06:11:38 crc kubenswrapper[4818]: I1122 06:11:38.444588 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-86757bfcb-s9bbk_34f2acb8-906b-4520-86c2-bb603a964489/horizon-log/0.log" Nov 22 06:11:38 crc kubenswrapper[4818]: I1122 06:11:38.447272 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-6h722_69bfd3ff-b747-43c0-82ab-ced23f3bc502/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:11:38 crc kubenswrapper[4818]: I1122 06:11:38.558971 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-fld7g_969421c2-6440-41e6-9866-c624dfe1cbae/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:11:38 crc kubenswrapper[4818]: I1122 06:11:38.771562 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29396521-h4m77_be3d9b50-d430-4a39-85e9-3afb138cfdf4/keystone-cron/0.log" Nov 22 06:11:38 crc kubenswrapper[4818]: I1122 06:11:38.892222 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_fd030ac8-aebe-4cde-b0e8-403f3049a922/kube-state-metrics/0.log" Nov 22 06:11:39 crc kubenswrapper[4818]: I1122 06:11:39.037912 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-595lb_806c9298-1aad-4e84-a80a-b2078f548d03/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:11:39 crc kubenswrapper[4818]: I1122 06:11:39.425630 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a/probe/0.log" Nov 22 06:11:39 crc kubenswrapper[4818]: I1122 06:11:39.594584 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_7feb3390-f51a-43f4-86c6-f44c30ebe2ab/manila-api/0.log" Nov 22 06:11:39 crc kubenswrapper[4818]: I1122 06:11:39.597990 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_fbdd2d2f-acd1-4f4d-8cf9-9878caa35c2a/manila-scheduler/0.log" Nov 22 06:11:39 crc kubenswrapper[4818]: I1122 06:11:39.788299 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_a5573538-84fb-4f11-9186-0dfd22a2232c/probe/0.log" Nov 22 06:11:39 crc kubenswrapper[4818]: I1122 06:11:39.842047 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-54d7c687fd-74zvh_fe562192-8025-4265-826d-6e7b1400b281/keystone-api/0.log" Nov 22 06:11:40 crc kubenswrapper[4818]: I1122 06:11:40.106408 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_7feb3390-f51a-43f4-86c6-f44c30ebe2ab/manila-api-log/0.log" Nov 22 06:11:40 crc kubenswrapper[4818]: I1122 06:11:40.229311 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-64754b546f-kqpj9_81f8a1cb-f2b5-429b-b12b-d2ff49b2e888/neutron-api/0.log" Nov 22 06:11:40 crc kubenswrapper[4818]: I1122 06:11:40.285724 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_a5573538-84fb-4f11-9186-0dfd22a2232c/manila-share/0.log" Nov 22 06:11:40 crc kubenswrapper[4818]: I1122 06:11:40.296083 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-64754b546f-kqpj9_81f8a1cb-f2b5-429b-b12b-d2ff49b2e888/neutron-httpd/0.log" Nov 22 06:11:40 crc kubenswrapper[4818]: I1122 06:11:40.425427 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-vz54z_e80824ae-495b-475f-a266-44879bb04e25/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:11:40 crc kubenswrapper[4818]: I1122 06:11:40.722901 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_ed42de57-a99e-4940-9874-49f6f703f8aa/nova-api-log/0.log" Nov 22 06:11:40 crc kubenswrapper[4818]: I1122 06:11:40.888043 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_5f15bdce-2f90-40df-a66e-141fb5d9f79f/nova-cell0-conductor-conductor/0.log" Nov 22 06:11:41 crc kubenswrapper[4818]: I1122 06:11:41.089677 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_f9d0ba71-b442-4c8a-87bd-72a8ca7284b2/nova-cell1-conductor-conductor/0.log" Nov 22 06:11:41 crc kubenswrapper[4818]: I1122 06:11:41.160632 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_ed42de57-a99e-4940-9874-49f6f703f8aa/nova-api-api/0.log" Nov 22 06:11:41 crc kubenswrapper[4818]: I1122 06:11:41.263006 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_3f5dcaf3-117e-4024-b1fc-41878a0eefd8/nova-cell1-novncproxy-novncproxy/0.log" Nov 22 06:11:41 crc kubenswrapper[4818]: I1122 06:11:41.358053 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8vm8q_38964063-2a2d-4728-9ccc-d5eca9db132d/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:11:41 crc kubenswrapper[4818]: I1122 06:11:41.590311 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e2b602e6-0bbf-4137-a806-d974b00003d7/nova-metadata-log/0.log" Nov 22 06:11:41 crc kubenswrapper[4818]: I1122 06:11:41.834628 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_f9271e30-e243-4016-bc57-2a10e514deb0/nova-scheduler-scheduler/0.log" Nov 22 06:11:41 crc kubenswrapper[4818]: I1122 06:11:41.879298 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_72cac26c-ba82-41c7-a405-a9162867abea/mysql-bootstrap/0.log" Nov 22 06:11:42 crc kubenswrapper[4818]: I1122 06:11:42.053914 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_72cac26c-ba82-41c7-a405-a9162867abea/mysql-bootstrap/0.log" Nov 22 06:11:42 crc kubenswrapper[4818]: I1122 06:11:42.103150 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_72cac26c-ba82-41c7-a405-a9162867abea/galera/0.log" Nov 22 06:11:42 crc kubenswrapper[4818]: I1122 06:11:42.409734 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_4714282b-39ba-4691-8fbd-c3e76120fbb5/mysql-bootstrap/0.log" Nov 22 06:11:42 crc kubenswrapper[4818]: I1122 06:11:42.748903 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_4714282b-39ba-4691-8fbd-c3e76120fbb5/mysql-bootstrap/0.log" Nov 22 06:11:42 crc kubenswrapper[4818]: I1122 06:11:42.943121 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_ac27ad40-d8e9-4ebb-9623-3136a1f9753c/openstackclient/0.log" Nov 22 06:11:42 crc kubenswrapper[4818]: I1122 06:11:42.950809 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_4714282b-39ba-4691-8fbd-c3e76120fbb5/galera/0.log" Nov 22 06:11:43 crc kubenswrapper[4818]: I1122 06:11:43.145974 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-jgq5d_7dec9ecf-e438-4334-a20e-757ce6fff445/ovn-controller/0.log" Nov 22 06:11:43 crc kubenswrapper[4818]: I1122 06:11:43.300748 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e2b602e6-0bbf-4137-a806-d974b00003d7/nova-metadata-metadata/0.log" Nov 22 06:11:43 crc kubenswrapper[4818]: I1122 06:11:43.356240 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-rsg7p_5b90b727-2867-4825-a5c1-a3549da28378/openstack-network-exporter/0.log" Nov 22 06:11:43 crc kubenswrapper[4818]: I1122 06:11:43.413311 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kpdh5_1b31a0d3-3489-430a-853a-40d8715f7e7f/ovsdb-server-init/0.log" Nov 22 06:11:43 crc kubenswrapper[4818]: I1122 06:11:43.570734 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kpdh5_1b31a0d3-3489-430a-853a-40d8715f7e7f/ovs-vswitchd/0.log" Nov 22 06:11:43 crc kubenswrapper[4818]: I1122 06:11:43.620348 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kpdh5_1b31a0d3-3489-430a-853a-40d8715f7e7f/ovsdb-server-init/0.log" Nov 22 06:11:43 crc kubenswrapper[4818]: I1122 06:11:43.633419 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kpdh5_1b31a0d3-3489-430a-853a-40d8715f7e7f/ovsdb-server/0.log" Nov 22 06:11:43 crc kubenswrapper[4818]: I1122 06:11:43.823427 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-k2rlm_0c86c5bf-d3d5-48b2-adfe-8da2c05d2cef/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:11:43 crc kubenswrapper[4818]: I1122 06:11:43.838912 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_6ab8b0e2-6df1-4a6b-b800-93e5b51ee430/openstack-network-exporter/0.log" Nov 22 06:11:44 crc kubenswrapper[4818]: I1122 06:11:44.383911 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_6ab8b0e2-6df1-4a6b-b800-93e5b51ee430/ovn-northd/0.log" Nov 22 06:11:44 crc kubenswrapper[4818]: I1122 06:11:44.426457 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_52d9e522-f84d-4671-9981-d08b3d7e0bbf/openstack-network-exporter/0.log" Nov 22 06:11:44 crc kubenswrapper[4818]: I1122 06:11:44.448676 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_52d9e522-f84d-4671-9981-d08b3d7e0bbf/ovsdbserver-nb/0.log" Nov 22 06:11:44 crc kubenswrapper[4818]: I1122 06:11:44.617657 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_5b0746b7-8400-44c8-b4c5-c6b6e3811e91/openstack-network-exporter/0.log" Nov 22 06:11:44 crc kubenswrapper[4818]: I1122 06:11:44.657855 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_5b0746b7-8400-44c8-b4c5-c6b6e3811e91/ovsdbserver-sb/0.log" Nov 22 06:11:44 crc kubenswrapper[4818]: I1122 06:11:44.892981 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7b464767cb-jcwc8_4e48ac54-5e91-4747-b9e9-28aeb21180fd/placement-api/0.log" Nov 22 06:11:44 crc kubenswrapper[4818]: I1122 06:11:44.941110 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7b464767cb-jcwc8_4e48ac54-5e91-4747-b9e9-28aeb21180fd/placement-log/0.log" Nov 22 06:11:44 crc kubenswrapper[4818]: I1122 06:11:44.965249 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3e1d70d4-6ada-47bf-a7f0-086ac6b891a0/setup-container/0.log" Nov 22 06:11:45 crc kubenswrapper[4818]: I1122 06:11:45.179987 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3e1d70d4-6ada-47bf-a7f0-086ac6b891a0/rabbitmq/0.log" Nov 22 06:11:45 crc kubenswrapper[4818]: I1122 06:11:45.186543 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86/setup-container/0.log" Nov 22 06:11:45 crc kubenswrapper[4818]: I1122 06:11:45.230819 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3e1d70d4-6ada-47bf-a7f0-086ac6b891a0/setup-container/0.log" Nov 22 06:11:45 crc kubenswrapper[4818]: I1122 06:11:45.421461 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86/rabbitmq/0.log" Nov 22 06:11:45 crc kubenswrapper[4818]: I1122 06:11:45.467192 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_70fd5462-fc4d-4e67-ace7-1f4e9a2a7c86/setup-container/0.log" Nov 22 06:11:45 crc kubenswrapper[4818]: I1122 06:11:45.474010 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-fgxxx_a072b598-ba0f-4b0d-8978-2d6e824ff5f6/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:11:45 crc kubenswrapper[4818]: I1122 06:11:45.701987 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-fvkst_0c466daa-c1c7-4d3c-a502-cfb8f424f6b2/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:11:45 crc kubenswrapper[4818]: I1122 06:11:45.748747 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-8b757_5d2cfc6e-9af5-4110-add7-b03c44ffbd33/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:11:45 crc kubenswrapper[4818]: I1122 06:11:45.934562 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-bg6f2_376ee003-816b-45fa-a57d-b9d2fa8120eb/ssh-known-hosts-edpm-deployment/0.log" Nov 22 06:11:46 crc kubenswrapper[4818]: I1122 06:11:46.054580 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_df264e71-c1de-4e1f-9a56-24f5bd284a80/tempest-tests-tempest-tests-runner/0.log" Nov 22 06:11:46 crc kubenswrapper[4818]: I1122 06:11:46.200592 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_fce70419-9c81-461c-bcb5-e829c1868e03/test-operator-logs-container/0.log" Nov 22 06:11:46 crc kubenswrapper[4818]: I1122 06:11:46.297698 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-2gvqx_68cb051d-9556-47e4-acc2-3bdb9b60ebdd/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 06:12:02 crc kubenswrapper[4818]: I1122 06:12:02.797380 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="4ea3ac76-3418-4b1c-8800-d57f554e50e3" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Nov 22 06:12:06 crc kubenswrapper[4818]: I1122 06:12:06.455958 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_e98d9180-5e2b-4b38-8fc0-844072603c2d/memcached/0.log" Nov 22 06:12:14 crc kubenswrapper[4818]: I1122 06:12:14.200991 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-8vknn_69877edd-8b97-4838-bc41-c087fc4deb7e/kube-rbac-proxy/0.log" Nov 22 06:12:14 crc kubenswrapper[4818]: I1122 06:12:14.234766 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-8vknn_69877edd-8b97-4838-bc41-c087fc4deb7e/manager/0.log" Nov 22 06:12:14 crc kubenswrapper[4818]: I1122 06:12:14.314838 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-b2lp5_27ec8b4f-e5ec-483a-ada8-6a766b843b73/kube-rbac-proxy/0.log" Nov 22 06:12:14 crc kubenswrapper[4818]: I1122 06:12:14.432303 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-b2lp5_27ec8b4f-e5ec-483a-ada8-6a766b843b73/manager/0.log" Nov 22 06:12:14 crc kubenswrapper[4818]: I1122 06:12:14.527588 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-jttdm_b9a85b40-dd76-45b2-bd41-ecb5be7387da/kube-rbac-proxy/0.log" Nov 22 06:12:14 crc kubenswrapper[4818]: I1122 06:12:14.529422 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-jttdm_b9a85b40-dd76-45b2-bd41-ecb5be7387da/manager/0.log" Nov 22 06:12:14 crc kubenswrapper[4818]: I1122 06:12:14.638802 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw_a5e6bbd8-bd14-4785-addd-67c8028cbd22/util/0.log" Nov 22 06:12:14 crc kubenswrapper[4818]: I1122 06:12:14.810479 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw_a5e6bbd8-bd14-4785-addd-67c8028cbd22/pull/0.log" Nov 22 06:12:14 crc kubenswrapper[4818]: I1122 06:12:14.825890 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw_a5e6bbd8-bd14-4785-addd-67c8028cbd22/pull/0.log" Nov 22 06:12:14 crc kubenswrapper[4818]: I1122 06:12:14.884274 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw_a5e6bbd8-bd14-4785-addd-67c8028cbd22/util/0.log" Nov 22 06:12:14 crc kubenswrapper[4818]: I1122 06:12:14.954729 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw_a5e6bbd8-bd14-4785-addd-67c8028cbd22/util/0.log" Nov 22 06:12:14 crc kubenswrapper[4818]: I1122 06:12:14.981443 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw_a5e6bbd8-bd14-4785-addd-67c8028cbd22/pull/0.log" Nov 22 06:12:15 crc kubenswrapper[4818]: I1122 06:12:15.014957 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f08e459c0d73d2ecea43b95eacc4d2511bfed70a37a1b316eaa6718cc3vbnpw_a5e6bbd8-bd14-4785-addd-67c8028cbd22/extract/0.log" Nov 22 06:12:15 crc kubenswrapper[4818]: I1122 06:12:15.122851 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-cl4tl_4d915797-cd52-4f07-9028-4f80411fe65b/kube-rbac-proxy/0.log" Nov 22 06:12:15 crc kubenswrapper[4818]: I1122 06:12:15.219413 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-cl4tl_4d915797-cd52-4f07-9028-4f80411fe65b/manager/0.log" Nov 22 06:12:15 crc kubenswrapper[4818]: I1122 06:12:15.322326 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-qt9nt_10da5c26-c5b3-4477-a4d4-a5c98741dfb1/kube-rbac-proxy/0.log" Nov 22 06:12:15 crc kubenswrapper[4818]: I1122 06:12:15.323375 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-qt9nt_10da5c26-c5b3-4477-a4d4-a5c98741dfb1/manager/0.log" Nov 22 06:12:15 crc kubenswrapper[4818]: I1122 06:12:15.433701 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-4zl42_05bab737-6cda-4891-a014-202ab05b141c/kube-rbac-proxy/0.log" Nov 22 06:12:15 crc kubenswrapper[4818]: I1122 06:12:15.470485 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-4zl42_05bab737-6cda-4891-a014-202ab05b141c/manager/0.log" Nov 22 06:12:15 crc kubenswrapper[4818]: I1122 06:12:15.618968 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7875d8bb94-dn8xx_7389ff62-006d-4d11-8237-e57c9279b0f3/kube-rbac-proxy/0.log" Nov 22 06:12:15 crc kubenswrapper[4818]: I1122 06:12:15.748952 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-ncfxv_c31258b3-6cac-41e8-a350-1ea7b360b848/kube-rbac-proxy/0.log" Nov 22 06:12:15 crc kubenswrapper[4818]: I1122 06:12:15.769388 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7875d8bb94-dn8xx_7389ff62-006d-4d11-8237-e57c9279b0f3/manager/0.log" Nov 22 06:12:15 crc kubenswrapper[4818]: I1122 06:12:15.822722 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-ncfxv_c31258b3-6cac-41e8-a350-1ea7b360b848/manager/0.log" Nov 22 06:12:15 crc kubenswrapper[4818]: I1122 06:12:15.913114 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-pqd4c_d6b59804-efc5-4f60-b9d3-bf0e998c8f89/kube-rbac-proxy/0.log" Nov 22 06:12:15 crc kubenswrapper[4818]: I1122 06:12:15.995270 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-pqd4c_d6b59804-efc5-4f60-b9d3-bf0e998c8f89/manager/0.log" Nov 22 06:12:16 crc kubenswrapper[4818]: I1122 06:12:16.115248 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-85bf8d6cb7-lzmmp_4838407d-e782-4aab-8fbf-861f3ba9019a/kube-rbac-proxy/0.log" Nov 22 06:12:16 crc kubenswrapper[4818]: I1122 06:12:16.151021 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-85bf8d6cb7-lzmmp_4838407d-e782-4aab-8fbf-861f3ba9019a/manager/0.log" Nov 22 06:12:16 crc kubenswrapper[4818]: I1122 06:12:16.268823 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-8wk6c_96a59329-8f96-43a5-9abc-fa8d1d7009fb/kube-rbac-proxy/0.log" Nov 22 06:12:16 crc kubenswrapper[4818]: I1122 06:12:16.335728 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-8wk6c_96a59329-8f96-43a5-9abc-fa8d1d7009fb/manager/0.log" Nov 22 06:12:16 crc kubenswrapper[4818]: I1122 06:12:16.383892 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-xtplb_68668b7f-5fc3-4774-8f0d-3eae71839d52/kube-rbac-proxy/0.log" Nov 22 06:12:16 crc kubenswrapper[4818]: I1122 06:12:16.486220 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-xtplb_68668b7f-5fc3-4774-8f0d-3eae71839d52/manager/0.log" Nov 22 06:12:16 crc kubenswrapper[4818]: I1122 06:12:16.543721 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-lx77w_531548a2-b8d3-4a2a-a6e9-1bb2af1891d8/kube-rbac-proxy/0.log" Nov 22 06:12:16 crc kubenswrapper[4818]: I1122 06:12:16.667501 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-lx77w_531548a2-b8d3-4a2a-a6e9-1bb2af1891d8/manager/0.log" Nov 22 06:12:16 crc kubenswrapper[4818]: I1122 06:12:16.715090 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-rf9vq_10d7f6df-ae52-4dab-b49d-764d95f77935/kube-rbac-proxy/0.log" Nov 22 06:12:16 crc kubenswrapper[4818]: I1122 06:12:16.778737 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-rf9vq_10d7f6df-ae52-4dab-b49d-764d95f77935/manager/0.log" Nov 22 06:12:16 crc kubenswrapper[4818]: I1122 06:12:16.853959 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-n54fq_9797ea62-91f5-4b4f-a9f0-27911cf629e4/kube-rbac-proxy/0.log" Nov 22 06:12:16 crc kubenswrapper[4818]: I1122 06:12:16.871414 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-n54fq_9797ea62-91f5-4b4f-a9f0-27911cf629e4/manager/0.log" Nov 22 06:12:17 crc kubenswrapper[4818]: I1122 06:12:17.012931 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6b57b64d88-pnqbd_44f3f34b-8fa2-479c-a876-947fab621708/kube-rbac-proxy/0.log" Nov 22 06:12:17 crc kubenswrapper[4818]: I1122 06:12:17.200664 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-59b954bb6d-6chhn_04e2d18d-98fc-4fa8-9a1f-1742f497d737/kube-rbac-proxy/0.log" Nov 22 06:12:17 crc kubenswrapper[4818]: I1122 06:12:17.381595 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-tn592_124c736f-9e78-422d-b317-b7263a45c55e/registry-server/0.log" Nov 22 06:12:17 crc kubenswrapper[4818]: I1122 06:12:17.440147 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-59b954bb6d-6chhn_04e2d18d-98fc-4fa8-9a1f-1742f497d737/operator/0.log" Nov 22 06:12:17 crc kubenswrapper[4818]: I1122 06:12:17.526936 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-8dhvf_06ec2990-cd45-4808-9748-2a1d75c5015f/kube-rbac-proxy/0.log" Nov 22 06:12:17 crc kubenswrapper[4818]: I1122 06:12:17.686563 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-56phb_786ab73e-012b-47f6-b033-9fa9aab6d6cb/kube-rbac-proxy/0.log" Nov 22 06:12:17 crc kubenswrapper[4818]: I1122 06:12:17.687604 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-8dhvf_06ec2990-cd45-4808-9748-2a1d75c5015f/manager/0.log" Nov 22 06:12:17 crc kubenswrapper[4818]: I1122 06:12:17.748695 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-56phb_786ab73e-012b-47f6-b033-9fa9aab6d6cb/manager/0.log" Nov 22 06:12:17 crc kubenswrapper[4818]: I1122 06:12:17.897601 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-d6957_62c4e1cf-72fb-4a9b-88a1-350c1e20f360/operator/0.log" Nov 22 06:12:17 crc kubenswrapper[4818]: I1122 06:12:17.949046 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-vkj22_f04a965b-f5f4-45af-843f-862498956853/kube-rbac-proxy/0.log" Nov 22 06:12:18 crc kubenswrapper[4818]: I1122 06:12:18.116934 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6d4bf84b58-cjphz_bba055bd-511b-4af6-bc94-4cadf7a84853/kube-rbac-proxy/0.log" Nov 22 06:12:18 crc kubenswrapper[4818]: I1122 06:12:18.125384 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-vkj22_f04a965b-f5f4-45af-843f-862498956853/manager/0.log" Nov 22 06:12:18 crc kubenswrapper[4818]: I1122 06:12:18.158340 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6b57b64d88-pnqbd_44f3f34b-8fa2-479c-a876-947fab621708/manager/0.log" Nov 22 06:12:18 crc kubenswrapper[4818]: I1122 06:12:18.261666 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6d4bf84b58-cjphz_bba055bd-511b-4af6-bc94-4cadf7a84853/manager/0.log" Nov 22 06:12:18 crc kubenswrapper[4818]: I1122 06:12:18.351543 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-pb45n_e342ff02-d6d9-4be0-b97f-87dea6653a61/manager/0.log" Nov 22 06:12:18 crc kubenswrapper[4818]: I1122 06:12:18.377754 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-pb45n_e342ff02-d6d9-4be0-b97f-87dea6653a61/kube-rbac-proxy/0.log" Nov 22 06:12:18 crc kubenswrapper[4818]: I1122 06:12:18.445136 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-sqvdq_2b80bb62-ec21-44c6-91e2-b422b9813255/kube-rbac-proxy/0.log" Nov 22 06:12:18 crc kubenswrapper[4818]: I1122 06:12:18.475697 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-sqvdq_2b80bb62-ec21-44c6-91e2-b422b9813255/manager/0.log" Nov 22 06:12:21 crc kubenswrapper[4818]: I1122 06:12:21.264813 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 06:12:21 crc kubenswrapper[4818]: I1122 06:12:21.266409 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 06:12:35 crc kubenswrapper[4818]: I1122 06:12:35.815854 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zftfz_7c429c72-1ae2-4b7d-9786-57ed9341bd48/control-plane-machine-set-operator/0.log" Nov 22 06:12:35 crc kubenswrapper[4818]: I1122 06:12:35.981339 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-8kzxl_8d3b850f-8917-46af-b713-891a7c979d2e/machine-api-operator/0.log" Nov 22 06:12:36 crc kubenswrapper[4818]: I1122 06:12:36.035358 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-8kzxl_8d3b850f-8917-46af-b713-891a7c979d2e/kube-rbac-proxy/0.log" Nov 22 06:12:51 crc kubenswrapper[4818]: I1122 06:12:51.264349 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 06:12:51 crc kubenswrapper[4818]: I1122 06:12:51.264913 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 06:12:52 crc kubenswrapper[4818]: I1122 06:12:52.075827 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5ndqt"] Nov 22 06:12:52 crc kubenswrapper[4818]: E1122 06:12:52.076706 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39def298-f711-4bcb-8fee-e95d7c37c606" containerName="container-00" Nov 22 06:12:52 crc kubenswrapper[4818]: I1122 06:12:52.076724 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="39def298-f711-4bcb-8fee-e95d7c37c606" containerName="container-00" Nov 22 06:12:52 crc kubenswrapper[4818]: I1122 06:12:52.077028 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="39def298-f711-4bcb-8fee-e95d7c37c606" containerName="container-00" Nov 22 06:12:52 crc kubenswrapper[4818]: I1122 06:12:52.078928 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5ndqt" Nov 22 06:12:52 crc kubenswrapper[4818]: I1122 06:12:52.110892 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5ndqt"] Nov 22 06:12:52 crc kubenswrapper[4818]: I1122 06:12:52.145388 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14280d97-ea1e-4f33-a2db-a6d92acf04af-utilities\") pod \"community-operators-5ndqt\" (UID: \"14280d97-ea1e-4f33-a2db-a6d92acf04af\") " pod="openshift-marketplace/community-operators-5ndqt" Nov 22 06:12:52 crc kubenswrapper[4818]: I1122 06:12:52.145440 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pwqb\" (UniqueName: \"kubernetes.io/projected/14280d97-ea1e-4f33-a2db-a6d92acf04af-kube-api-access-4pwqb\") pod \"community-operators-5ndqt\" (UID: \"14280d97-ea1e-4f33-a2db-a6d92acf04af\") " pod="openshift-marketplace/community-operators-5ndqt" Nov 22 06:12:52 crc kubenswrapper[4818]: I1122 06:12:52.145476 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14280d97-ea1e-4f33-a2db-a6d92acf04af-catalog-content\") pod \"community-operators-5ndqt\" (UID: \"14280d97-ea1e-4f33-a2db-a6d92acf04af\") " pod="openshift-marketplace/community-operators-5ndqt" Nov 22 06:12:52 crc kubenswrapper[4818]: I1122 06:12:52.237131 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-9l45k_c62c123f-00fc-4a5e-9649-87121c5f793a/cert-manager-controller/0.log" Nov 22 06:12:52 crc kubenswrapper[4818]: I1122 06:12:52.246894 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14280d97-ea1e-4f33-a2db-a6d92acf04af-utilities\") pod \"community-operators-5ndqt\" (UID: \"14280d97-ea1e-4f33-a2db-a6d92acf04af\") " pod="openshift-marketplace/community-operators-5ndqt" Nov 22 06:12:52 crc kubenswrapper[4818]: I1122 06:12:52.246939 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pwqb\" (UniqueName: \"kubernetes.io/projected/14280d97-ea1e-4f33-a2db-a6d92acf04af-kube-api-access-4pwqb\") pod \"community-operators-5ndqt\" (UID: \"14280d97-ea1e-4f33-a2db-a6d92acf04af\") " pod="openshift-marketplace/community-operators-5ndqt" Nov 22 06:12:52 crc kubenswrapper[4818]: I1122 06:12:52.246965 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14280d97-ea1e-4f33-a2db-a6d92acf04af-catalog-content\") pod \"community-operators-5ndqt\" (UID: \"14280d97-ea1e-4f33-a2db-a6d92acf04af\") " pod="openshift-marketplace/community-operators-5ndqt" Nov 22 06:12:52 crc kubenswrapper[4818]: I1122 06:12:52.247472 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14280d97-ea1e-4f33-a2db-a6d92acf04af-catalog-content\") pod \"community-operators-5ndqt\" (UID: \"14280d97-ea1e-4f33-a2db-a6d92acf04af\") " pod="openshift-marketplace/community-operators-5ndqt" Nov 22 06:12:52 crc kubenswrapper[4818]: I1122 06:12:52.247544 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14280d97-ea1e-4f33-a2db-a6d92acf04af-utilities\") pod \"community-operators-5ndqt\" (UID: \"14280d97-ea1e-4f33-a2db-a6d92acf04af\") " pod="openshift-marketplace/community-operators-5ndqt" Nov 22 06:12:52 crc kubenswrapper[4818]: I1122 06:12:52.263973 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pwqb\" (UniqueName: \"kubernetes.io/projected/14280d97-ea1e-4f33-a2db-a6d92acf04af-kube-api-access-4pwqb\") pod \"community-operators-5ndqt\" (UID: \"14280d97-ea1e-4f33-a2db-a6d92acf04af\") " pod="openshift-marketplace/community-operators-5ndqt" Nov 22 06:12:52 crc kubenswrapper[4818]: I1122 06:12:52.399538 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5ndqt" Nov 22 06:12:52 crc kubenswrapper[4818]: I1122 06:12:52.557377 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-248l2_de576ad7-b4e6-4047-92d1-dc19f477be33/cert-manager-cainjector/0.log" Nov 22 06:12:52 crc kubenswrapper[4818]: I1122 06:12:52.733618 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-4bmxf_77efc3b2-d4e5-426e-ada4-de8c900575f9/cert-manager-webhook/0.log" Nov 22 06:12:53 crc kubenswrapper[4818]: I1122 06:12:53.023450 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5ndqt"] Nov 22 06:12:53 crc kubenswrapper[4818]: W1122 06:12:53.026995 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14280d97_ea1e_4f33_a2db_a6d92acf04af.slice/crio-10c6b2b07fcaacae9e98eb1e82755afbddb81c6aaa45e910eb3f948235552efa WatchSource:0}: Error finding container 10c6b2b07fcaacae9e98eb1e82755afbddb81c6aaa45e910eb3f948235552efa: Status 404 returned error can't find the container with id 10c6b2b07fcaacae9e98eb1e82755afbddb81c6aaa45e910eb3f948235552efa Nov 22 06:12:53 crc kubenswrapper[4818]: I1122 06:12:53.909563 4818 generic.go:334] "Generic (PLEG): container finished" podID="14280d97-ea1e-4f33-a2db-a6d92acf04af" containerID="8189a4a09e1fe02f6f111fc0c815b856d88d778026365ad8c2d14f4a4bdcee89" exitCode=0 Nov 22 06:12:53 crc kubenswrapper[4818]: I1122 06:12:53.909604 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5ndqt" event={"ID":"14280d97-ea1e-4f33-a2db-a6d92acf04af","Type":"ContainerDied","Data":"8189a4a09e1fe02f6f111fc0c815b856d88d778026365ad8c2d14f4a4bdcee89"} Nov 22 06:12:53 crc kubenswrapper[4818]: I1122 06:12:53.909861 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5ndqt" event={"ID":"14280d97-ea1e-4f33-a2db-a6d92acf04af","Type":"ContainerStarted","Data":"10c6b2b07fcaacae9e98eb1e82755afbddb81c6aaa45e910eb3f948235552efa"} Nov 22 06:12:53 crc kubenswrapper[4818]: I1122 06:12:53.912311 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 06:12:54 crc kubenswrapper[4818]: I1122 06:12:54.921385 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5ndqt" event={"ID":"14280d97-ea1e-4f33-a2db-a6d92acf04af","Type":"ContainerStarted","Data":"f35c8ac33f3a7b44553a6c327792a49f5bc97437a5eb04eb1c2492741c06cca5"} Nov 22 06:12:55 crc kubenswrapper[4818]: I1122 06:12:55.932400 4818 generic.go:334] "Generic (PLEG): container finished" podID="14280d97-ea1e-4f33-a2db-a6d92acf04af" containerID="f35c8ac33f3a7b44553a6c327792a49f5bc97437a5eb04eb1c2492741c06cca5" exitCode=0 Nov 22 06:12:55 crc kubenswrapper[4818]: I1122 06:12:55.932500 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5ndqt" event={"ID":"14280d97-ea1e-4f33-a2db-a6d92acf04af","Type":"ContainerDied","Data":"f35c8ac33f3a7b44553a6c327792a49f5bc97437a5eb04eb1c2492741c06cca5"} Nov 22 06:12:57 crc kubenswrapper[4818]: I1122 06:12:57.954998 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5ndqt" event={"ID":"14280d97-ea1e-4f33-a2db-a6d92acf04af","Type":"ContainerStarted","Data":"558d09bafe776f9e7142b5b625cd88e6f202e229d9e1351397939a52c9b0b8fb"} Nov 22 06:12:57 crc kubenswrapper[4818]: I1122 06:12:57.994955 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5ndqt" podStartSLOduration=3.085138189 podStartE2EDuration="5.994938056s" podCreationTimestamp="2025-11-22 06:12:52 +0000 UTC" firstStartedPulling="2025-11-22 06:12:53.912045086 +0000 UTC m=+5126.486461613" lastFinishedPulling="2025-11-22 06:12:56.821844933 +0000 UTC m=+5129.396261480" observedRunningTime="2025-11-22 06:12:57.982095153 +0000 UTC m=+5130.556511680" watchObservedRunningTime="2025-11-22 06:12:57.994938056 +0000 UTC m=+5130.569354583" Nov 22 06:13:02 crc kubenswrapper[4818]: I1122 06:13:02.400547 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5ndqt" Nov 22 06:13:02 crc kubenswrapper[4818]: I1122 06:13:02.401225 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5ndqt" Nov 22 06:13:02 crc kubenswrapper[4818]: I1122 06:13:02.451606 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5ndqt" Nov 22 06:13:03 crc kubenswrapper[4818]: I1122 06:13:03.058058 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5ndqt" Nov 22 06:13:03 crc kubenswrapper[4818]: I1122 06:13:03.115978 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5ndqt"] Nov 22 06:13:05 crc kubenswrapper[4818]: I1122 06:13:05.026988 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5ndqt" podUID="14280d97-ea1e-4f33-a2db-a6d92acf04af" containerName="registry-server" containerID="cri-o://558d09bafe776f9e7142b5b625cd88e6f202e229d9e1351397939a52c9b0b8fb" gracePeriod=2 Nov 22 06:13:05 crc kubenswrapper[4818]: I1122 06:13:05.580402 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5ndqt" Nov 22 06:13:05 crc kubenswrapper[4818]: I1122 06:13:05.644529 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14280d97-ea1e-4f33-a2db-a6d92acf04af-catalog-content\") pod \"14280d97-ea1e-4f33-a2db-a6d92acf04af\" (UID: \"14280d97-ea1e-4f33-a2db-a6d92acf04af\") " Nov 22 06:13:05 crc kubenswrapper[4818]: I1122 06:13:05.644662 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pwqb\" (UniqueName: \"kubernetes.io/projected/14280d97-ea1e-4f33-a2db-a6d92acf04af-kube-api-access-4pwqb\") pod \"14280d97-ea1e-4f33-a2db-a6d92acf04af\" (UID: \"14280d97-ea1e-4f33-a2db-a6d92acf04af\") " Nov 22 06:13:05 crc kubenswrapper[4818]: I1122 06:13:05.644738 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14280d97-ea1e-4f33-a2db-a6d92acf04af-utilities\") pod \"14280d97-ea1e-4f33-a2db-a6d92acf04af\" (UID: \"14280d97-ea1e-4f33-a2db-a6d92acf04af\") " Nov 22 06:13:05 crc kubenswrapper[4818]: I1122 06:13:05.645682 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14280d97-ea1e-4f33-a2db-a6d92acf04af-utilities" (OuterVolumeSpecName: "utilities") pod "14280d97-ea1e-4f33-a2db-a6d92acf04af" (UID: "14280d97-ea1e-4f33-a2db-a6d92acf04af"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 06:13:05 crc kubenswrapper[4818]: I1122 06:13:05.665451 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14280d97-ea1e-4f33-a2db-a6d92acf04af-kube-api-access-4pwqb" (OuterVolumeSpecName: "kube-api-access-4pwqb") pod "14280d97-ea1e-4f33-a2db-a6d92acf04af" (UID: "14280d97-ea1e-4f33-a2db-a6d92acf04af"). InnerVolumeSpecName "kube-api-access-4pwqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 06:13:05 crc kubenswrapper[4818]: I1122 06:13:05.694168 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14280d97-ea1e-4f33-a2db-a6d92acf04af-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "14280d97-ea1e-4f33-a2db-a6d92acf04af" (UID: "14280d97-ea1e-4f33-a2db-a6d92acf04af"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 06:13:05 crc kubenswrapper[4818]: I1122 06:13:05.746996 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14280d97-ea1e-4f33-a2db-a6d92acf04af-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 06:13:05 crc kubenswrapper[4818]: I1122 06:13:05.747033 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pwqb\" (UniqueName: \"kubernetes.io/projected/14280d97-ea1e-4f33-a2db-a6d92acf04af-kube-api-access-4pwqb\") on node \"crc\" DevicePath \"\"" Nov 22 06:13:05 crc kubenswrapper[4818]: I1122 06:13:05.747044 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14280d97-ea1e-4f33-a2db-a6d92acf04af-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 06:13:06 crc kubenswrapper[4818]: I1122 06:13:06.039639 4818 generic.go:334] "Generic (PLEG): container finished" podID="14280d97-ea1e-4f33-a2db-a6d92acf04af" containerID="558d09bafe776f9e7142b5b625cd88e6f202e229d9e1351397939a52c9b0b8fb" exitCode=0 Nov 22 06:13:06 crc kubenswrapper[4818]: I1122 06:13:06.039945 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5ndqt" Nov 22 06:13:06 crc kubenswrapper[4818]: I1122 06:13:06.039836 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5ndqt" event={"ID":"14280d97-ea1e-4f33-a2db-a6d92acf04af","Type":"ContainerDied","Data":"558d09bafe776f9e7142b5b625cd88e6f202e229d9e1351397939a52c9b0b8fb"} Nov 22 06:13:06 crc kubenswrapper[4818]: I1122 06:13:06.040241 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5ndqt" event={"ID":"14280d97-ea1e-4f33-a2db-a6d92acf04af","Type":"ContainerDied","Data":"10c6b2b07fcaacae9e98eb1e82755afbddb81c6aaa45e910eb3f948235552efa"} Nov 22 06:13:06 crc kubenswrapper[4818]: I1122 06:13:06.040292 4818 scope.go:117] "RemoveContainer" containerID="558d09bafe776f9e7142b5b625cd88e6f202e229d9e1351397939a52c9b0b8fb" Nov 22 06:13:06 crc kubenswrapper[4818]: I1122 06:13:06.059882 4818 scope.go:117] "RemoveContainer" containerID="f35c8ac33f3a7b44553a6c327792a49f5bc97437a5eb04eb1c2492741c06cca5" Nov 22 06:13:06 crc kubenswrapper[4818]: I1122 06:13:06.091419 4818 scope.go:117] "RemoveContainer" containerID="8189a4a09e1fe02f6f111fc0c815b856d88d778026365ad8c2d14f4a4bdcee89" Nov 22 06:13:06 crc kubenswrapper[4818]: I1122 06:13:06.091833 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5ndqt"] Nov 22 06:13:06 crc kubenswrapper[4818]: I1122 06:13:06.103634 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5ndqt"] Nov 22 06:13:06 crc kubenswrapper[4818]: I1122 06:13:06.127529 4818 scope.go:117] "RemoveContainer" containerID="558d09bafe776f9e7142b5b625cd88e6f202e229d9e1351397939a52c9b0b8fb" Nov 22 06:13:06 crc kubenswrapper[4818]: E1122 06:13:06.127979 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"558d09bafe776f9e7142b5b625cd88e6f202e229d9e1351397939a52c9b0b8fb\": container with ID starting with 558d09bafe776f9e7142b5b625cd88e6f202e229d9e1351397939a52c9b0b8fb not found: ID does not exist" containerID="558d09bafe776f9e7142b5b625cd88e6f202e229d9e1351397939a52c9b0b8fb" Nov 22 06:13:06 crc kubenswrapper[4818]: I1122 06:13:06.128008 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"558d09bafe776f9e7142b5b625cd88e6f202e229d9e1351397939a52c9b0b8fb"} err="failed to get container status \"558d09bafe776f9e7142b5b625cd88e6f202e229d9e1351397939a52c9b0b8fb\": rpc error: code = NotFound desc = could not find container \"558d09bafe776f9e7142b5b625cd88e6f202e229d9e1351397939a52c9b0b8fb\": container with ID starting with 558d09bafe776f9e7142b5b625cd88e6f202e229d9e1351397939a52c9b0b8fb not found: ID does not exist" Nov 22 06:13:06 crc kubenswrapper[4818]: I1122 06:13:06.128028 4818 scope.go:117] "RemoveContainer" containerID="f35c8ac33f3a7b44553a6c327792a49f5bc97437a5eb04eb1c2492741c06cca5" Nov 22 06:13:06 crc kubenswrapper[4818]: E1122 06:13:06.128325 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f35c8ac33f3a7b44553a6c327792a49f5bc97437a5eb04eb1c2492741c06cca5\": container with ID starting with f35c8ac33f3a7b44553a6c327792a49f5bc97437a5eb04eb1c2492741c06cca5 not found: ID does not exist" containerID="f35c8ac33f3a7b44553a6c327792a49f5bc97437a5eb04eb1c2492741c06cca5" Nov 22 06:13:06 crc kubenswrapper[4818]: I1122 06:13:06.128348 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f35c8ac33f3a7b44553a6c327792a49f5bc97437a5eb04eb1c2492741c06cca5"} err="failed to get container status \"f35c8ac33f3a7b44553a6c327792a49f5bc97437a5eb04eb1c2492741c06cca5\": rpc error: code = NotFound desc = could not find container \"f35c8ac33f3a7b44553a6c327792a49f5bc97437a5eb04eb1c2492741c06cca5\": container with ID starting with f35c8ac33f3a7b44553a6c327792a49f5bc97437a5eb04eb1c2492741c06cca5 not found: ID does not exist" Nov 22 06:13:06 crc kubenswrapper[4818]: I1122 06:13:06.128363 4818 scope.go:117] "RemoveContainer" containerID="8189a4a09e1fe02f6f111fc0c815b856d88d778026365ad8c2d14f4a4bdcee89" Nov 22 06:13:06 crc kubenswrapper[4818]: E1122 06:13:06.128595 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8189a4a09e1fe02f6f111fc0c815b856d88d778026365ad8c2d14f4a4bdcee89\": container with ID starting with 8189a4a09e1fe02f6f111fc0c815b856d88d778026365ad8c2d14f4a4bdcee89 not found: ID does not exist" containerID="8189a4a09e1fe02f6f111fc0c815b856d88d778026365ad8c2d14f4a4bdcee89" Nov 22 06:13:06 crc kubenswrapper[4818]: I1122 06:13:06.128613 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8189a4a09e1fe02f6f111fc0c815b856d88d778026365ad8c2d14f4a4bdcee89"} err="failed to get container status \"8189a4a09e1fe02f6f111fc0c815b856d88d778026365ad8c2d14f4a4bdcee89\": rpc error: code = NotFound desc = could not find container \"8189a4a09e1fe02f6f111fc0c815b856d88d778026365ad8c2d14f4a4bdcee89\": container with ID starting with 8189a4a09e1fe02f6f111fc0c815b856d88d778026365ad8c2d14f4a4bdcee89 not found: ID does not exist" Nov 22 06:13:06 crc kubenswrapper[4818]: I1122 06:13:06.306463 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14280d97-ea1e-4f33-a2db-a6d92acf04af" path="/var/lib/kubelet/pods/14280d97-ea1e-4f33-a2db-a6d92acf04af/volumes" Nov 22 06:13:06 crc kubenswrapper[4818]: I1122 06:13:06.877937 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-d2f5x_614393c8-1594-4694-8281-a1e20cab9209/nmstate-console-plugin/0.log" Nov 22 06:13:07 crc kubenswrapper[4818]: I1122 06:13:07.049969 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-ftnnk_695d974e-caa1-43c9-b619-363891b2bb7c/nmstate-handler/0.log" Nov 22 06:13:07 crc kubenswrapper[4818]: I1122 06:13:07.063928 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-f8v82_38740f9a-864c-44b3-8f95-f9ffecc8e9d5/kube-rbac-proxy/0.log" Nov 22 06:13:07 crc kubenswrapper[4818]: I1122 06:13:07.103760 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-f8v82_38740f9a-864c-44b3-8f95-f9ffecc8e9d5/nmstate-metrics/0.log" Nov 22 06:13:07 crc kubenswrapper[4818]: I1122 06:13:07.226272 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-r2wjf_05f359e3-f133-443a-9067-148ddc6fac3b/nmstate-operator/0.log" Nov 22 06:13:07 crc kubenswrapper[4818]: I1122 06:13:07.516884 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-d94tc_b0f0e1e5-3f4d-463b-8c00-91db5867fdfc/nmstate-webhook/0.log" Nov 22 06:13:21 crc kubenswrapper[4818]: I1122 06:13:21.264450 4818 patch_prober.go:28] interesting pod/machine-config-daemon-5rjvw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 06:13:21 crc kubenswrapper[4818]: I1122 06:13:21.264897 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 06:13:21 crc kubenswrapper[4818]: I1122 06:13:21.264945 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" Nov 22 06:13:21 crc kubenswrapper[4818]: I1122 06:13:21.265733 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa"} pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 06:13:21 crc kubenswrapper[4818]: I1122 06:13:21.265786 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" containerName="machine-config-daemon" containerID="cri-o://07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" gracePeriod=600 Nov 22 06:13:21 crc kubenswrapper[4818]: E1122 06:13:21.407366 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:13:21 crc kubenswrapper[4818]: I1122 06:13:21.837725 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-wbzs6_0fdd1082-4473-4604-8ced-2294bf4ba248/kube-rbac-proxy/0.log" Nov 22 06:13:21 crc kubenswrapper[4818]: I1122 06:13:21.974384 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-wbzs6_0fdd1082-4473-4604-8ced-2294bf4ba248/controller/0.log" Nov 22 06:13:22 crc kubenswrapper[4818]: I1122 06:13:22.087864 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-frr-files/0.log" Nov 22 06:13:22 crc kubenswrapper[4818]: I1122 06:13:22.174998 4818 generic.go:334] "Generic (PLEG): container finished" podID="0971c9ec-29ce-44eb-9875-ad4664143305" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" exitCode=0 Nov 22 06:13:22 crc kubenswrapper[4818]: I1122 06:13:22.175038 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerDied","Data":"07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa"} Nov 22 06:13:22 crc kubenswrapper[4818]: I1122 06:13:22.175070 4818 scope.go:117] "RemoveContainer" containerID="d2100269659d0a3f462da37bed1df37938f3da7494989b3a814f1fdfb9d5a0f9" Nov 22 06:13:22 crc kubenswrapper[4818]: I1122 06:13:22.175728 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:13:22 crc kubenswrapper[4818]: E1122 06:13:22.175972 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:13:22 crc kubenswrapper[4818]: I1122 06:13:22.201485 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-reloader/0.log" Nov 22 06:13:22 crc kubenswrapper[4818]: I1122 06:13:22.213187 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-frr-files/0.log" Nov 22 06:13:22 crc kubenswrapper[4818]: I1122 06:13:22.272173 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-metrics/0.log" Nov 22 06:13:22 crc kubenswrapper[4818]: I1122 06:13:22.279245 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-reloader/0.log" Nov 22 06:13:22 crc kubenswrapper[4818]: I1122 06:13:22.496080 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-reloader/0.log" Nov 22 06:13:22 crc kubenswrapper[4818]: I1122 06:13:22.514389 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-metrics/0.log" Nov 22 06:13:22 crc kubenswrapper[4818]: I1122 06:13:22.533242 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-metrics/0.log" Nov 22 06:13:22 crc kubenswrapper[4818]: I1122 06:13:22.539220 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-frr-files/0.log" Nov 22 06:13:22 crc kubenswrapper[4818]: I1122 06:13:22.677467 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-frr-files/0.log" Nov 22 06:13:22 crc kubenswrapper[4818]: I1122 06:13:22.678452 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-reloader/0.log" Nov 22 06:13:22 crc kubenswrapper[4818]: I1122 06:13:22.699846 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/cp-metrics/0.log" Nov 22 06:13:22 crc kubenswrapper[4818]: I1122 06:13:22.709189 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/controller/0.log" Nov 22 06:13:22 crc kubenswrapper[4818]: I1122 06:13:22.844588 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/kube-rbac-proxy/0.log" Nov 22 06:13:22 crc kubenswrapper[4818]: I1122 06:13:22.875266 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/kube-rbac-proxy-frr/0.log" Nov 22 06:13:22 crc kubenswrapper[4818]: I1122 06:13:22.884209 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/frr-metrics/0.log" Nov 22 06:13:23 crc kubenswrapper[4818]: I1122 06:13:23.052178 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/reloader/0.log" Nov 22 06:13:23 crc kubenswrapper[4818]: I1122 06:13:23.107451 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-rss4n_a90765d9-f4ec-4f76-921b-f8fe52fa2cdc/frr-k8s-webhook-server/0.log" Nov 22 06:13:23 crc kubenswrapper[4818]: I1122 06:13:23.385416 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5b7cbb7cc5-ff54p_254a82ef-b456-4309-93fb-a236494bf3d4/manager/0.log" Nov 22 06:13:23 crc kubenswrapper[4818]: I1122 06:13:23.506379 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-757d8876f9-6qjkv_63241779-80e0-43a3-bb9c-6530a4a6ae05/webhook-server/0.log" Nov 22 06:13:23 crc kubenswrapper[4818]: I1122 06:13:23.610753 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8mwxx_dbc8ab54-ce5e-4320-8deb-4d597cb4f82a/kube-rbac-proxy/0.log" Nov 22 06:13:24 crc kubenswrapper[4818]: I1122 06:13:24.150761 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8mwxx_dbc8ab54-ce5e-4320-8deb-4d597cb4f82a/speaker/0.log" Nov 22 06:13:24 crc kubenswrapper[4818]: I1122 06:13:24.527002 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ld8qn_4bf0b266-290e-4c4f-8ddb-17c38155751e/frr/0.log" Nov 22 06:13:36 crc kubenswrapper[4818]: I1122 06:13:36.291189 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:13:36 crc kubenswrapper[4818]: E1122 06:13:36.291970 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:13:38 crc kubenswrapper[4818]: I1122 06:13:38.576470 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd_cc4bf9a2-88ba-4882-9908-f1e2f27c82ac/util/0.log" Nov 22 06:13:38 crc kubenswrapper[4818]: I1122 06:13:38.689417 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd_cc4bf9a2-88ba-4882-9908-f1e2f27c82ac/util/0.log" Nov 22 06:13:38 crc kubenswrapper[4818]: I1122 06:13:38.693227 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd_cc4bf9a2-88ba-4882-9908-f1e2f27c82ac/pull/0.log" Nov 22 06:13:38 crc kubenswrapper[4818]: I1122 06:13:38.751858 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd_cc4bf9a2-88ba-4882-9908-f1e2f27c82ac/pull/0.log" Nov 22 06:13:39 crc kubenswrapper[4818]: I1122 06:13:39.495793 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd_cc4bf9a2-88ba-4882-9908-f1e2f27c82ac/util/0.log" Nov 22 06:13:39 crc kubenswrapper[4818]: I1122 06:13:39.502057 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd_cc4bf9a2-88ba-4882-9908-f1e2f27c82ac/pull/0.log" Nov 22 06:13:39 crc kubenswrapper[4818]: I1122 06:13:39.552679 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772esvcwd_cc4bf9a2-88ba-4882-9908-f1e2f27c82ac/extract/0.log" Nov 22 06:13:39 crc kubenswrapper[4818]: I1122 06:13:39.695072 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcnvz_603888a5-b8ab-4378-ab59-d044221e403c/extract-utilities/0.log" Nov 22 06:13:39 crc kubenswrapper[4818]: I1122 06:13:39.839905 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcnvz_603888a5-b8ab-4378-ab59-d044221e403c/extract-utilities/0.log" Nov 22 06:13:39 crc kubenswrapper[4818]: I1122 06:13:39.856324 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcnvz_603888a5-b8ab-4378-ab59-d044221e403c/extract-content/0.log" Nov 22 06:13:39 crc kubenswrapper[4818]: I1122 06:13:39.877167 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcnvz_603888a5-b8ab-4378-ab59-d044221e403c/extract-content/0.log" Nov 22 06:13:40 crc kubenswrapper[4818]: I1122 06:13:40.004581 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcnvz_603888a5-b8ab-4378-ab59-d044221e403c/extract-utilities/0.log" Nov 22 06:13:40 crc kubenswrapper[4818]: I1122 06:13:40.043531 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcnvz_603888a5-b8ab-4378-ab59-d044221e403c/extract-content/0.log" Nov 22 06:13:40 crc kubenswrapper[4818]: I1122 06:13:40.241852 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cdjr2_fd603ec5-ddaf-4d51-a4fc-09442ae6692e/extract-utilities/0.log" Nov 22 06:13:40 crc kubenswrapper[4818]: I1122 06:13:40.459644 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cdjr2_fd603ec5-ddaf-4d51-a4fc-09442ae6692e/extract-content/0.log" Nov 22 06:13:40 crc kubenswrapper[4818]: I1122 06:13:40.555437 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cdjr2_fd603ec5-ddaf-4d51-a4fc-09442ae6692e/extract-utilities/0.log" Nov 22 06:13:40 crc kubenswrapper[4818]: I1122 06:13:40.619480 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cdjr2_fd603ec5-ddaf-4d51-a4fc-09442ae6692e/extract-content/0.log" Nov 22 06:13:40 crc kubenswrapper[4818]: I1122 06:13:40.632894 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcnvz_603888a5-b8ab-4378-ab59-d044221e403c/registry-server/0.log" Nov 22 06:13:40 crc kubenswrapper[4818]: I1122 06:13:40.693373 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cdjr2_fd603ec5-ddaf-4d51-a4fc-09442ae6692e/extract-utilities/0.log" Nov 22 06:13:40 crc kubenswrapper[4818]: I1122 06:13:40.730391 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cdjr2_fd603ec5-ddaf-4d51-a4fc-09442ae6692e/extract-content/0.log" Nov 22 06:13:40 crc kubenswrapper[4818]: I1122 06:13:40.896830 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d_cd479145-6121-48b1-b54a-b122e24a03fd/util/0.log" Nov 22 06:13:41 crc kubenswrapper[4818]: I1122 06:13:41.068477 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cdjr2_fd603ec5-ddaf-4d51-a4fc-09442ae6692e/registry-server/0.log" Nov 22 06:13:41 crc kubenswrapper[4818]: I1122 06:13:41.107279 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d_cd479145-6121-48b1-b54a-b122e24a03fd/util/0.log" Nov 22 06:13:41 crc kubenswrapper[4818]: I1122 06:13:41.167861 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d_cd479145-6121-48b1-b54a-b122e24a03fd/pull/0.log" Nov 22 06:13:41 crc kubenswrapper[4818]: I1122 06:13:41.210728 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d_cd479145-6121-48b1-b54a-b122e24a03fd/pull/0.log" Nov 22 06:13:41 crc kubenswrapper[4818]: I1122 06:13:41.393795 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d_cd479145-6121-48b1-b54a-b122e24a03fd/util/0.log" Nov 22 06:13:41 crc kubenswrapper[4818]: I1122 06:13:41.413885 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d_cd479145-6121-48b1-b54a-b122e24a03fd/pull/0.log" Nov 22 06:13:41 crc kubenswrapper[4818]: I1122 06:13:41.418710 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-cxqxn_6e9cbedb-a168-495f-8746-ccf21c4e330c/marketplace-operator/0.log" Nov 22 06:13:41 crc kubenswrapper[4818]: I1122 06:13:41.424000 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c685h7d_cd479145-6121-48b1-b54a-b122e24a03fd/extract/0.log" Nov 22 06:13:41 crc kubenswrapper[4818]: I1122 06:13:41.593483 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rcbjp_d83d4bfc-8842-4d00-96d9-779a89080441/extract-utilities/0.log" Nov 22 06:13:41 crc kubenswrapper[4818]: I1122 06:13:41.752545 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rcbjp_d83d4bfc-8842-4d00-96d9-779a89080441/extract-utilities/0.log" Nov 22 06:13:41 crc kubenswrapper[4818]: I1122 06:13:41.781628 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rcbjp_d83d4bfc-8842-4d00-96d9-779a89080441/extract-content/0.log" Nov 22 06:13:41 crc kubenswrapper[4818]: I1122 06:13:41.789422 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rcbjp_d83d4bfc-8842-4d00-96d9-779a89080441/extract-content/0.log" Nov 22 06:13:41 crc kubenswrapper[4818]: I1122 06:13:41.962848 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rcbjp_d83d4bfc-8842-4d00-96d9-779a89080441/extract-utilities/0.log" Nov 22 06:13:42 crc kubenswrapper[4818]: I1122 06:13:42.016072 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rcbjp_d83d4bfc-8842-4d00-96d9-779a89080441/extract-content/0.log" Nov 22 06:13:42 crc kubenswrapper[4818]: I1122 06:13:42.038590 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8lq9_759a3ee8-f990-4960-bc61-9893ca8cbc6f/extract-utilities/0.log" Nov 22 06:13:42 crc kubenswrapper[4818]: I1122 06:13:42.177472 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rcbjp_d83d4bfc-8842-4d00-96d9-779a89080441/registry-server/0.log" Nov 22 06:13:42 crc kubenswrapper[4818]: I1122 06:13:42.251079 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8lq9_759a3ee8-f990-4960-bc61-9893ca8cbc6f/extract-utilities/0.log" Nov 22 06:13:42 crc kubenswrapper[4818]: I1122 06:13:42.276992 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8lq9_759a3ee8-f990-4960-bc61-9893ca8cbc6f/extract-content/0.log" Nov 22 06:13:42 crc kubenswrapper[4818]: I1122 06:13:42.300325 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8lq9_759a3ee8-f990-4960-bc61-9893ca8cbc6f/extract-content/0.log" Nov 22 06:13:42 crc kubenswrapper[4818]: I1122 06:13:42.473500 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8lq9_759a3ee8-f990-4960-bc61-9893ca8cbc6f/extract-utilities/0.log" Nov 22 06:13:42 crc kubenswrapper[4818]: I1122 06:13:42.476428 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8lq9_759a3ee8-f990-4960-bc61-9893ca8cbc6f/extract-content/0.log" Nov 22 06:13:43 crc kubenswrapper[4818]: I1122 06:13:43.178222 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8lq9_759a3ee8-f990-4960-bc61-9893ca8cbc6f/registry-server/0.log" Nov 22 06:13:49 crc kubenswrapper[4818]: I1122 06:13:49.291405 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:13:49 crc kubenswrapper[4818]: E1122 06:13:49.293057 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:14:03 crc kubenswrapper[4818]: I1122 06:14:03.291288 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:14:03 crc kubenswrapper[4818]: E1122 06:14:03.292149 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:14:16 crc kubenswrapper[4818]: I1122 06:14:16.291291 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:14:16 crc kubenswrapper[4818]: E1122 06:14:16.291872 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:14:31 crc kubenswrapper[4818]: I1122 06:14:31.292017 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:14:31 crc kubenswrapper[4818]: E1122 06:14:31.293982 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:14:37 crc kubenswrapper[4818]: I1122 06:14:37.998676 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5d6vl"] Nov 22 06:14:38 crc kubenswrapper[4818]: E1122 06:14:37.999652 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14280d97-ea1e-4f33-a2db-a6d92acf04af" containerName="registry-server" Nov 22 06:14:38 crc kubenswrapper[4818]: I1122 06:14:37.999668 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="14280d97-ea1e-4f33-a2db-a6d92acf04af" containerName="registry-server" Nov 22 06:14:38 crc kubenswrapper[4818]: E1122 06:14:37.999711 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14280d97-ea1e-4f33-a2db-a6d92acf04af" containerName="extract-content" Nov 22 06:14:38 crc kubenswrapper[4818]: I1122 06:14:37.999719 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="14280d97-ea1e-4f33-a2db-a6d92acf04af" containerName="extract-content" Nov 22 06:14:38 crc kubenswrapper[4818]: E1122 06:14:37.999734 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14280d97-ea1e-4f33-a2db-a6d92acf04af" containerName="extract-utilities" Nov 22 06:14:38 crc kubenswrapper[4818]: I1122 06:14:37.999743 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="14280d97-ea1e-4f33-a2db-a6d92acf04af" containerName="extract-utilities" Nov 22 06:14:38 crc kubenswrapper[4818]: I1122 06:14:37.999952 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="14280d97-ea1e-4f33-a2db-a6d92acf04af" containerName="registry-server" Nov 22 06:14:38 crc kubenswrapper[4818]: I1122 06:14:38.001796 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5d6vl" Nov 22 06:14:38 crc kubenswrapper[4818]: I1122 06:14:38.020466 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5d6vl"] Nov 22 06:14:38 crc kubenswrapper[4818]: I1122 06:14:38.114290 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc0ca77c-a97b-467f-9c4d-a79612defabb-utilities\") pod \"redhat-operators-5d6vl\" (UID: \"cc0ca77c-a97b-467f-9c4d-a79612defabb\") " pod="openshift-marketplace/redhat-operators-5d6vl" Nov 22 06:14:38 crc kubenswrapper[4818]: I1122 06:14:38.114371 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc0ca77c-a97b-467f-9c4d-a79612defabb-catalog-content\") pod \"redhat-operators-5d6vl\" (UID: \"cc0ca77c-a97b-467f-9c4d-a79612defabb\") " pod="openshift-marketplace/redhat-operators-5d6vl" Nov 22 06:14:38 crc kubenswrapper[4818]: I1122 06:14:38.114422 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx2q5\" (UniqueName: \"kubernetes.io/projected/cc0ca77c-a97b-467f-9c4d-a79612defabb-kube-api-access-xx2q5\") pod \"redhat-operators-5d6vl\" (UID: \"cc0ca77c-a97b-467f-9c4d-a79612defabb\") " pod="openshift-marketplace/redhat-operators-5d6vl" Nov 22 06:14:38 crc kubenswrapper[4818]: I1122 06:14:38.216546 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx2q5\" (UniqueName: \"kubernetes.io/projected/cc0ca77c-a97b-467f-9c4d-a79612defabb-kube-api-access-xx2q5\") pod \"redhat-operators-5d6vl\" (UID: \"cc0ca77c-a97b-467f-9c4d-a79612defabb\") " pod="openshift-marketplace/redhat-operators-5d6vl" Nov 22 06:14:38 crc kubenswrapper[4818]: I1122 06:14:38.216832 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc0ca77c-a97b-467f-9c4d-a79612defabb-utilities\") pod \"redhat-operators-5d6vl\" (UID: \"cc0ca77c-a97b-467f-9c4d-a79612defabb\") " pod="openshift-marketplace/redhat-operators-5d6vl" Nov 22 06:14:38 crc kubenswrapper[4818]: I1122 06:14:38.216979 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc0ca77c-a97b-467f-9c4d-a79612defabb-catalog-content\") pod \"redhat-operators-5d6vl\" (UID: \"cc0ca77c-a97b-467f-9c4d-a79612defabb\") " pod="openshift-marketplace/redhat-operators-5d6vl" Nov 22 06:14:38 crc kubenswrapper[4818]: I1122 06:14:38.217411 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc0ca77c-a97b-467f-9c4d-a79612defabb-utilities\") pod \"redhat-operators-5d6vl\" (UID: \"cc0ca77c-a97b-467f-9c4d-a79612defabb\") " pod="openshift-marketplace/redhat-operators-5d6vl" Nov 22 06:14:38 crc kubenswrapper[4818]: I1122 06:14:38.217725 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc0ca77c-a97b-467f-9c4d-a79612defabb-catalog-content\") pod \"redhat-operators-5d6vl\" (UID: \"cc0ca77c-a97b-467f-9c4d-a79612defabb\") " pod="openshift-marketplace/redhat-operators-5d6vl" Nov 22 06:14:38 crc kubenswrapper[4818]: I1122 06:14:38.238897 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx2q5\" (UniqueName: \"kubernetes.io/projected/cc0ca77c-a97b-467f-9c4d-a79612defabb-kube-api-access-xx2q5\") pod \"redhat-operators-5d6vl\" (UID: \"cc0ca77c-a97b-467f-9c4d-a79612defabb\") " pod="openshift-marketplace/redhat-operators-5d6vl" Nov 22 06:14:38 crc kubenswrapper[4818]: I1122 06:14:38.326470 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5d6vl" Nov 22 06:14:38 crc kubenswrapper[4818]: I1122 06:14:38.859161 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5d6vl"] Nov 22 06:14:38 crc kubenswrapper[4818]: I1122 06:14:38.949995 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5d6vl" event={"ID":"cc0ca77c-a97b-467f-9c4d-a79612defabb","Type":"ContainerStarted","Data":"03795426b6334ed3a2c7fa6a857af63b754d5b0d8af188d48c523f36e123e21a"} Nov 22 06:14:39 crc kubenswrapper[4818]: I1122 06:14:39.969662 4818 generic.go:334] "Generic (PLEG): container finished" podID="cc0ca77c-a97b-467f-9c4d-a79612defabb" containerID="243e3762077932c2805120f1242b9f1b4d5320382422a40367fa2498fbe9180d" exitCode=0 Nov 22 06:14:39 crc kubenswrapper[4818]: I1122 06:14:39.969731 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5d6vl" event={"ID":"cc0ca77c-a97b-467f-9c4d-a79612defabb","Type":"ContainerDied","Data":"243e3762077932c2805120f1242b9f1b4d5320382422a40367fa2498fbe9180d"} Nov 22 06:14:43 crc kubenswrapper[4818]: I1122 06:14:43.018381 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5d6vl" event={"ID":"cc0ca77c-a97b-467f-9c4d-a79612defabb","Type":"ContainerStarted","Data":"ebb33c8797859b3ccf70d9cfd8960a93b77799046fe6b9bf04662b436565e067"} Nov 22 06:14:46 crc kubenswrapper[4818]: I1122 06:14:46.292522 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:14:46 crc kubenswrapper[4818]: E1122 06:14:46.293349 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:14:48 crc kubenswrapper[4818]: I1122 06:14:48.094718 4818 generic.go:334] "Generic (PLEG): container finished" podID="cc0ca77c-a97b-467f-9c4d-a79612defabb" containerID="ebb33c8797859b3ccf70d9cfd8960a93b77799046fe6b9bf04662b436565e067" exitCode=0 Nov 22 06:14:48 crc kubenswrapper[4818]: I1122 06:14:48.094853 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5d6vl" event={"ID":"cc0ca77c-a97b-467f-9c4d-a79612defabb","Type":"ContainerDied","Data":"ebb33c8797859b3ccf70d9cfd8960a93b77799046fe6b9bf04662b436565e067"} Nov 22 06:14:51 crc kubenswrapper[4818]: I1122 06:14:51.125810 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5d6vl" event={"ID":"cc0ca77c-a97b-467f-9c4d-a79612defabb","Type":"ContainerStarted","Data":"5840d293e9db920bb9d1c7d35d1a13b31837568d4e3998c872becf21653c6761"} Nov 22 06:14:51 crc kubenswrapper[4818]: I1122 06:14:51.156521 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5d6vl" podStartSLOduration=4.469668185 podStartE2EDuration="14.156495925s" podCreationTimestamp="2025-11-22 06:14:37 +0000 UTC" firstStartedPulling="2025-11-22 06:14:39.972682927 +0000 UTC m=+5232.547099464" lastFinishedPulling="2025-11-22 06:14:49.659510667 +0000 UTC m=+5242.233927204" observedRunningTime="2025-11-22 06:14:51.147684998 +0000 UTC m=+5243.722101515" watchObservedRunningTime="2025-11-22 06:14:51.156495925 +0000 UTC m=+5243.730912462" Nov 22 06:14:58 crc kubenswrapper[4818]: I1122 06:14:58.326972 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5d6vl" Nov 22 06:14:58 crc kubenswrapper[4818]: I1122 06:14:58.327648 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5d6vl" Nov 22 06:14:58 crc kubenswrapper[4818]: I1122 06:14:58.430559 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5d6vl" Nov 22 06:14:59 crc kubenswrapper[4818]: I1122 06:14:59.291977 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:14:59 crc kubenswrapper[4818]: E1122 06:14:59.292919 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:14:59 crc kubenswrapper[4818]: I1122 06:14:59.295039 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5d6vl" Nov 22 06:14:59 crc kubenswrapper[4818]: I1122 06:14:59.376473 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5d6vl"] Nov 22 06:15:00 crc kubenswrapper[4818]: I1122 06:15:00.189273 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w"] Nov 22 06:15:00 crc kubenswrapper[4818]: I1122 06:15:00.190522 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w" Nov 22 06:15:00 crc kubenswrapper[4818]: I1122 06:15:00.193017 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 06:15:00 crc kubenswrapper[4818]: I1122 06:15:00.193121 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 06:15:00 crc kubenswrapper[4818]: I1122 06:15:00.202069 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w"] Nov 22 06:15:00 crc kubenswrapper[4818]: I1122 06:15:00.265117 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncfz5\" (UniqueName: \"kubernetes.io/projected/77eeb152-32d3-48f7-ba18-5b510b0bdd5c-kube-api-access-ncfz5\") pod \"collect-profiles-29396535-txk9w\" (UID: \"77eeb152-32d3-48f7-ba18-5b510b0bdd5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w" Nov 22 06:15:00 crc kubenswrapper[4818]: I1122 06:15:00.265202 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/77eeb152-32d3-48f7-ba18-5b510b0bdd5c-secret-volume\") pod \"collect-profiles-29396535-txk9w\" (UID: \"77eeb152-32d3-48f7-ba18-5b510b0bdd5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w" Nov 22 06:15:00 crc kubenswrapper[4818]: I1122 06:15:00.265236 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/77eeb152-32d3-48f7-ba18-5b510b0bdd5c-config-volume\") pod \"collect-profiles-29396535-txk9w\" (UID: \"77eeb152-32d3-48f7-ba18-5b510b0bdd5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w" Nov 22 06:15:00 crc kubenswrapper[4818]: I1122 06:15:00.367144 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/77eeb152-32d3-48f7-ba18-5b510b0bdd5c-secret-volume\") pod \"collect-profiles-29396535-txk9w\" (UID: \"77eeb152-32d3-48f7-ba18-5b510b0bdd5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w" Nov 22 06:15:00 crc kubenswrapper[4818]: I1122 06:15:00.367232 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/77eeb152-32d3-48f7-ba18-5b510b0bdd5c-config-volume\") pod \"collect-profiles-29396535-txk9w\" (UID: \"77eeb152-32d3-48f7-ba18-5b510b0bdd5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w" Nov 22 06:15:00 crc kubenswrapper[4818]: I1122 06:15:00.367603 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncfz5\" (UniqueName: \"kubernetes.io/projected/77eeb152-32d3-48f7-ba18-5b510b0bdd5c-kube-api-access-ncfz5\") pod \"collect-profiles-29396535-txk9w\" (UID: \"77eeb152-32d3-48f7-ba18-5b510b0bdd5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w" Nov 22 06:15:00 crc kubenswrapper[4818]: I1122 06:15:00.368163 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/77eeb152-32d3-48f7-ba18-5b510b0bdd5c-config-volume\") pod \"collect-profiles-29396535-txk9w\" (UID: \"77eeb152-32d3-48f7-ba18-5b510b0bdd5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w" Nov 22 06:15:00 crc kubenswrapper[4818]: I1122 06:15:00.695902 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncfz5\" (UniqueName: \"kubernetes.io/projected/77eeb152-32d3-48f7-ba18-5b510b0bdd5c-kube-api-access-ncfz5\") pod \"collect-profiles-29396535-txk9w\" (UID: \"77eeb152-32d3-48f7-ba18-5b510b0bdd5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w" Nov 22 06:15:00 crc kubenswrapper[4818]: I1122 06:15:00.696386 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/77eeb152-32d3-48f7-ba18-5b510b0bdd5c-secret-volume\") pod \"collect-profiles-29396535-txk9w\" (UID: \"77eeb152-32d3-48f7-ba18-5b510b0bdd5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w" Nov 22 06:15:00 crc kubenswrapper[4818]: I1122 06:15:00.834756 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w" Nov 22 06:15:01 crc kubenswrapper[4818]: I1122 06:15:01.227362 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5d6vl" podUID="cc0ca77c-a97b-467f-9c4d-a79612defabb" containerName="registry-server" containerID="cri-o://5840d293e9db920bb9d1c7d35d1a13b31837568d4e3998c872becf21653c6761" gracePeriod=2 Nov 22 06:15:01 crc kubenswrapper[4818]: I1122 06:15:01.353007 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w"] Nov 22 06:15:02 crc kubenswrapper[4818]: I1122 06:15:02.242901 4818 generic.go:334] "Generic (PLEG): container finished" podID="cc0ca77c-a97b-467f-9c4d-a79612defabb" containerID="5840d293e9db920bb9d1c7d35d1a13b31837568d4e3998c872becf21653c6761" exitCode=0 Nov 22 06:15:02 crc kubenswrapper[4818]: I1122 06:15:02.243036 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5d6vl" event={"ID":"cc0ca77c-a97b-467f-9c4d-a79612defabb","Type":"ContainerDied","Data":"5840d293e9db920bb9d1c7d35d1a13b31837568d4e3998c872becf21653c6761"} Nov 22 06:15:02 crc kubenswrapper[4818]: I1122 06:15:02.245214 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w" event={"ID":"77eeb152-32d3-48f7-ba18-5b510b0bdd5c","Type":"ContainerStarted","Data":"1e96c9374a2783abcebaa9135072c166244c543093b1749bb4cb9c42e6c3e443"} Nov 22 06:15:02 crc kubenswrapper[4818]: I1122 06:15:02.245401 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w" event={"ID":"77eeb152-32d3-48f7-ba18-5b510b0bdd5c","Type":"ContainerStarted","Data":"c25b0818feca8a261e240372351883988739d08c11c00a49c016edcaa2a85446"} Nov 22 06:15:02 crc kubenswrapper[4818]: I1122 06:15:02.274220 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w" podStartSLOduration=2.274198623 podStartE2EDuration="2.274198623s" podCreationTimestamp="2025-11-22 06:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 06:15:02.270473664 +0000 UTC m=+5254.844890241" watchObservedRunningTime="2025-11-22 06:15:02.274198623 +0000 UTC m=+5254.848615160" Nov 22 06:15:03 crc kubenswrapper[4818]: I1122 06:15:03.039250 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5d6vl" Nov 22 06:15:03 crc kubenswrapper[4818]: I1122 06:15:03.233189 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc0ca77c-a97b-467f-9c4d-a79612defabb-catalog-content\") pod \"cc0ca77c-a97b-467f-9c4d-a79612defabb\" (UID: \"cc0ca77c-a97b-467f-9c4d-a79612defabb\") " Nov 22 06:15:03 crc kubenswrapper[4818]: I1122 06:15:03.233892 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc0ca77c-a97b-467f-9c4d-a79612defabb-utilities\") pod \"cc0ca77c-a97b-467f-9c4d-a79612defabb\" (UID: \"cc0ca77c-a97b-467f-9c4d-a79612defabb\") " Nov 22 06:15:03 crc kubenswrapper[4818]: I1122 06:15:03.234143 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xx2q5\" (UniqueName: \"kubernetes.io/projected/cc0ca77c-a97b-467f-9c4d-a79612defabb-kube-api-access-xx2q5\") pod \"cc0ca77c-a97b-467f-9c4d-a79612defabb\" (UID: \"cc0ca77c-a97b-467f-9c4d-a79612defabb\") " Nov 22 06:15:03 crc kubenswrapper[4818]: I1122 06:15:03.244191 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc0ca77c-a97b-467f-9c4d-a79612defabb-utilities" (OuterVolumeSpecName: "utilities") pod "cc0ca77c-a97b-467f-9c4d-a79612defabb" (UID: "cc0ca77c-a97b-467f-9c4d-a79612defabb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 06:15:03 crc kubenswrapper[4818]: I1122 06:15:03.244615 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc0ca77c-a97b-467f-9c4d-a79612defabb-kube-api-access-xx2q5" (OuterVolumeSpecName: "kube-api-access-xx2q5") pod "cc0ca77c-a97b-467f-9c4d-a79612defabb" (UID: "cc0ca77c-a97b-467f-9c4d-a79612defabb"). InnerVolumeSpecName "kube-api-access-xx2q5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 06:15:03 crc kubenswrapper[4818]: I1122 06:15:03.257158 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5d6vl" event={"ID":"cc0ca77c-a97b-467f-9c4d-a79612defabb","Type":"ContainerDied","Data":"03795426b6334ed3a2c7fa6a857af63b754d5b0d8af188d48c523f36e123e21a"} Nov 22 06:15:03 crc kubenswrapper[4818]: I1122 06:15:03.257275 4818 scope.go:117] "RemoveContainer" containerID="5840d293e9db920bb9d1c7d35d1a13b31837568d4e3998c872becf21653c6761" Nov 22 06:15:03 crc kubenswrapper[4818]: I1122 06:15:03.258198 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5d6vl" Nov 22 06:15:03 crc kubenswrapper[4818]: I1122 06:15:03.259636 4818 generic.go:334] "Generic (PLEG): container finished" podID="77eeb152-32d3-48f7-ba18-5b510b0bdd5c" containerID="1e96c9374a2783abcebaa9135072c166244c543093b1749bb4cb9c42e6c3e443" exitCode=0 Nov 22 06:15:03 crc kubenswrapper[4818]: I1122 06:15:03.259673 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w" event={"ID":"77eeb152-32d3-48f7-ba18-5b510b0bdd5c","Type":"ContainerDied","Data":"1e96c9374a2783abcebaa9135072c166244c543093b1749bb4cb9c42e6c3e443"} Nov 22 06:15:03 crc kubenswrapper[4818]: I1122 06:15:03.303824 4818 scope.go:117] "RemoveContainer" containerID="ebb33c8797859b3ccf70d9cfd8960a93b77799046fe6b9bf04662b436565e067" Nov 22 06:15:03 crc kubenswrapper[4818]: I1122 06:15:03.322467 4818 scope.go:117] "RemoveContainer" containerID="243e3762077932c2805120f1242b9f1b4d5320382422a40367fa2498fbe9180d" Nov 22 06:15:03 crc kubenswrapper[4818]: I1122 06:15:03.337515 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc0ca77c-a97b-467f-9c4d-a79612defabb-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 06:15:03 crc kubenswrapper[4818]: I1122 06:15:03.337546 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xx2q5\" (UniqueName: \"kubernetes.io/projected/cc0ca77c-a97b-467f-9c4d-a79612defabb-kube-api-access-xx2q5\") on node \"crc\" DevicePath \"\"" Nov 22 06:15:03 crc kubenswrapper[4818]: I1122 06:15:03.339783 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc0ca77c-a97b-467f-9c4d-a79612defabb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc0ca77c-a97b-467f-9c4d-a79612defabb" (UID: "cc0ca77c-a97b-467f-9c4d-a79612defabb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 06:15:03 crc kubenswrapper[4818]: I1122 06:15:03.439903 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc0ca77c-a97b-467f-9c4d-a79612defabb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 06:15:03 crc kubenswrapper[4818]: I1122 06:15:03.596528 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5d6vl"] Nov 22 06:15:03 crc kubenswrapper[4818]: I1122 06:15:03.608951 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5d6vl"] Nov 22 06:15:04 crc kubenswrapper[4818]: I1122 06:15:04.311531 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc0ca77c-a97b-467f-9c4d-a79612defabb" path="/var/lib/kubelet/pods/cc0ca77c-a97b-467f-9c4d-a79612defabb/volumes" Nov 22 06:15:04 crc kubenswrapper[4818]: I1122 06:15:04.663845 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w" Nov 22 06:15:04 crc kubenswrapper[4818]: I1122 06:15:04.765969 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77eeb152-32d3-48f7-ba18-5b510b0bdd5c-config-volume" (OuterVolumeSpecName: "config-volume") pod "77eeb152-32d3-48f7-ba18-5b510b0bdd5c" (UID: "77eeb152-32d3-48f7-ba18-5b510b0bdd5c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 06:15:04 crc kubenswrapper[4818]: I1122 06:15:04.766030 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/77eeb152-32d3-48f7-ba18-5b510b0bdd5c-config-volume\") pod \"77eeb152-32d3-48f7-ba18-5b510b0bdd5c\" (UID: \"77eeb152-32d3-48f7-ba18-5b510b0bdd5c\") " Nov 22 06:15:04 crc kubenswrapper[4818]: I1122 06:15:04.766311 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/77eeb152-32d3-48f7-ba18-5b510b0bdd5c-secret-volume\") pod \"77eeb152-32d3-48f7-ba18-5b510b0bdd5c\" (UID: \"77eeb152-32d3-48f7-ba18-5b510b0bdd5c\") " Nov 22 06:15:04 crc kubenswrapper[4818]: I1122 06:15:04.767371 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncfz5\" (UniqueName: \"kubernetes.io/projected/77eeb152-32d3-48f7-ba18-5b510b0bdd5c-kube-api-access-ncfz5\") pod \"77eeb152-32d3-48f7-ba18-5b510b0bdd5c\" (UID: \"77eeb152-32d3-48f7-ba18-5b510b0bdd5c\") " Nov 22 06:15:04 crc kubenswrapper[4818]: I1122 06:15:04.768925 4818 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/77eeb152-32d3-48f7-ba18-5b510b0bdd5c-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 06:15:04 crc kubenswrapper[4818]: I1122 06:15:04.785608 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77eeb152-32d3-48f7-ba18-5b510b0bdd5c-kube-api-access-ncfz5" (OuterVolumeSpecName: "kube-api-access-ncfz5") pod "77eeb152-32d3-48f7-ba18-5b510b0bdd5c" (UID: "77eeb152-32d3-48f7-ba18-5b510b0bdd5c"). InnerVolumeSpecName "kube-api-access-ncfz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 06:15:04 crc kubenswrapper[4818]: I1122 06:15:04.785609 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77eeb152-32d3-48f7-ba18-5b510b0bdd5c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "77eeb152-32d3-48f7-ba18-5b510b0bdd5c" (UID: "77eeb152-32d3-48f7-ba18-5b510b0bdd5c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 06:15:04 crc kubenswrapper[4818]: I1122 06:15:04.870563 4818 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/77eeb152-32d3-48f7-ba18-5b510b0bdd5c-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 06:15:04 crc kubenswrapper[4818]: I1122 06:15:04.870603 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncfz5\" (UniqueName: \"kubernetes.io/projected/77eeb152-32d3-48f7-ba18-5b510b0bdd5c-kube-api-access-ncfz5\") on node \"crc\" DevicePath \"\"" Nov 22 06:15:05 crc kubenswrapper[4818]: I1122 06:15:05.286598 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w" event={"ID":"77eeb152-32d3-48f7-ba18-5b510b0bdd5c","Type":"ContainerDied","Data":"c25b0818feca8a261e240372351883988739d08c11c00a49c016edcaa2a85446"} Nov 22 06:15:05 crc kubenswrapper[4818]: I1122 06:15:05.286705 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c25b0818feca8a261e240372351883988739d08c11c00a49c016edcaa2a85446" Nov 22 06:15:05 crc kubenswrapper[4818]: I1122 06:15:05.286643 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396535-txk9w" Nov 22 06:15:05 crc kubenswrapper[4818]: I1122 06:15:05.775588 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls"] Nov 22 06:15:05 crc kubenswrapper[4818]: I1122 06:15:05.786495 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396490-2ksls"] Nov 22 06:15:06 crc kubenswrapper[4818]: I1122 06:15:06.304969 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c0d4c12-ebd6-4020-aa27-d5dfda947b44" path="/var/lib/kubelet/pods/5c0d4c12-ebd6-4020-aa27-d5dfda947b44/volumes" Nov 22 06:15:14 crc kubenswrapper[4818]: I1122 06:15:14.291720 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:15:14 crc kubenswrapper[4818]: E1122 06:15:14.292768 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:15:25 crc kubenswrapper[4818]: I1122 06:15:25.290947 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:15:25 crc kubenswrapper[4818]: E1122 06:15:25.291980 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:15:32 crc kubenswrapper[4818]: I1122 06:15:32.607568 4818 generic.go:334] "Generic (PLEG): container finished" podID="cde37c03-b657-46b8-9a24-6ed81eac8c50" containerID="976cfeb045a0f62146acce89fab7e0d55deb6a063823564ae8c401f996757811" exitCode=0 Nov 22 06:15:32 crc kubenswrapper[4818]: I1122 06:15:32.607675 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bksd9/must-gather-nlvnh" event={"ID":"cde37c03-b657-46b8-9a24-6ed81eac8c50","Type":"ContainerDied","Data":"976cfeb045a0f62146acce89fab7e0d55deb6a063823564ae8c401f996757811"} Nov 22 06:15:32 crc kubenswrapper[4818]: I1122 06:15:32.608994 4818 scope.go:117] "RemoveContainer" containerID="976cfeb045a0f62146acce89fab7e0d55deb6a063823564ae8c401f996757811" Nov 22 06:15:32 crc kubenswrapper[4818]: I1122 06:15:32.795139 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bksd9_must-gather-nlvnh_cde37c03-b657-46b8-9a24-6ed81eac8c50/gather/0.log" Nov 22 06:15:40 crc kubenswrapper[4818]: I1122 06:15:40.291910 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:15:40 crc kubenswrapper[4818]: E1122 06:15:40.293345 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:15:43 crc kubenswrapper[4818]: I1122 06:15:43.432315 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bksd9/must-gather-nlvnh"] Nov 22 06:15:43 crc kubenswrapper[4818]: I1122 06:15:43.433004 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-bksd9/must-gather-nlvnh" podUID="cde37c03-b657-46b8-9a24-6ed81eac8c50" containerName="copy" containerID="cri-o://e56317ded1280e57b749879ff256808bebaa3a20ae0c363fb055b08ce17c79a1" gracePeriod=2 Nov 22 06:15:43 crc kubenswrapper[4818]: I1122 06:15:43.449735 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bksd9/must-gather-nlvnh"] Nov 22 06:15:43 crc kubenswrapper[4818]: I1122 06:15:43.755117 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bksd9_must-gather-nlvnh_cde37c03-b657-46b8-9a24-6ed81eac8c50/copy/0.log" Nov 22 06:15:43 crc kubenswrapper[4818]: I1122 06:15:43.755869 4818 generic.go:334] "Generic (PLEG): container finished" podID="cde37c03-b657-46b8-9a24-6ed81eac8c50" containerID="e56317ded1280e57b749879ff256808bebaa3a20ae0c363fb055b08ce17c79a1" exitCode=143 Nov 22 06:15:43 crc kubenswrapper[4818]: I1122 06:15:43.860699 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bksd9_must-gather-nlvnh_cde37c03-b657-46b8-9a24-6ed81eac8c50/copy/0.log" Nov 22 06:15:43 crc kubenswrapper[4818]: I1122 06:15:43.861356 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bksd9/must-gather-nlvnh" Nov 22 06:15:43 crc kubenswrapper[4818]: I1122 06:15:43.933621 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cde37c03-b657-46b8-9a24-6ed81eac8c50-must-gather-output\") pod \"cde37c03-b657-46b8-9a24-6ed81eac8c50\" (UID: \"cde37c03-b657-46b8-9a24-6ed81eac8c50\") " Nov 22 06:15:43 crc kubenswrapper[4818]: I1122 06:15:43.933807 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lj9ct\" (UniqueName: \"kubernetes.io/projected/cde37c03-b657-46b8-9a24-6ed81eac8c50-kube-api-access-lj9ct\") pod \"cde37c03-b657-46b8-9a24-6ed81eac8c50\" (UID: \"cde37c03-b657-46b8-9a24-6ed81eac8c50\") " Nov 22 06:15:43 crc kubenswrapper[4818]: I1122 06:15:43.942048 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cde37c03-b657-46b8-9a24-6ed81eac8c50-kube-api-access-lj9ct" (OuterVolumeSpecName: "kube-api-access-lj9ct") pod "cde37c03-b657-46b8-9a24-6ed81eac8c50" (UID: "cde37c03-b657-46b8-9a24-6ed81eac8c50"). InnerVolumeSpecName "kube-api-access-lj9ct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 06:15:44 crc kubenswrapper[4818]: I1122 06:15:44.036507 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lj9ct\" (UniqueName: \"kubernetes.io/projected/cde37c03-b657-46b8-9a24-6ed81eac8c50-kube-api-access-lj9ct\") on node \"crc\" DevicePath \"\"" Nov 22 06:15:44 crc kubenswrapper[4818]: I1122 06:15:44.109271 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cde37c03-b657-46b8-9a24-6ed81eac8c50-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "cde37c03-b657-46b8-9a24-6ed81eac8c50" (UID: "cde37c03-b657-46b8-9a24-6ed81eac8c50"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 06:15:44 crc kubenswrapper[4818]: I1122 06:15:44.138737 4818 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cde37c03-b657-46b8-9a24-6ed81eac8c50-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 22 06:15:44 crc kubenswrapper[4818]: I1122 06:15:44.306240 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cde37c03-b657-46b8-9a24-6ed81eac8c50" path="/var/lib/kubelet/pods/cde37c03-b657-46b8-9a24-6ed81eac8c50/volumes" Nov 22 06:15:44 crc kubenswrapper[4818]: I1122 06:15:44.774346 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bksd9_must-gather-nlvnh_cde37c03-b657-46b8-9a24-6ed81eac8c50/copy/0.log" Nov 22 06:15:44 crc kubenswrapper[4818]: I1122 06:15:44.774932 4818 scope.go:117] "RemoveContainer" containerID="e56317ded1280e57b749879ff256808bebaa3a20ae0c363fb055b08ce17c79a1" Nov 22 06:15:44 crc kubenswrapper[4818]: I1122 06:15:44.775009 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bksd9/must-gather-nlvnh" Nov 22 06:15:44 crc kubenswrapper[4818]: I1122 06:15:44.808974 4818 scope.go:117] "RemoveContainer" containerID="976cfeb045a0f62146acce89fab7e0d55deb6a063823564ae8c401f996757811" Nov 22 06:15:45 crc kubenswrapper[4818]: I1122 06:15:45.238585 4818 scope.go:117] "RemoveContainer" containerID="c414d1cd4492199df9c403e7ff11723287375fc6caa6b7abd4b200bdd18ecc2d" Nov 22 06:15:51 crc kubenswrapper[4818]: I1122 06:15:51.377689 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:15:51 crc kubenswrapper[4818]: E1122 06:15:51.378567 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:16:06 crc kubenswrapper[4818]: I1122 06:16:06.291405 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:16:06 crc kubenswrapper[4818]: E1122 06:16:06.292213 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:16:19 crc kubenswrapper[4818]: I1122 06:16:19.292086 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:16:19 crc kubenswrapper[4818]: E1122 06:16:19.293337 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:16:33 crc kubenswrapper[4818]: I1122 06:16:33.292056 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:16:33 crc kubenswrapper[4818]: E1122 06:16:33.293130 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:16:45 crc kubenswrapper[4818]: I1122 06:16:45.292878 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:16:45 crc kubenswrapper[4818]: E1122 06:16:45.294067 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.350185 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-drjjm"] Nov 22 06:16:55 crc kubenswrapper[4818]: E1122 06:16:55.352249 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cde37c03-b657-46b8-9a24-6ed81eac8c50" containerName="gather" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.352396 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="cde37c03-b657-46b8-9a24-6ed81eac8c50" containerName="gather" Nov 22 06:16:55 crc kubenswrapper[4818]: E1122 06:16:55.352412 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc0ca77c-a97b-467f-9c4d-a79612defabb" containerName="registry-server" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.352424 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc0ca77c-a97b-467f-9c4d-a79612defabb" containerName="registry-server" Nov 22 06:16:55 crc kubenswrapper[4818]: E1122 06:16:55.352450 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc0ca77c-a97b-467f-9c4d-a79612defabb" containerName="extract-utilities" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.352464 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc0ca77c-a97b-467f-9c4d-a79612defabb" containerName="extract-utilities" Nov 22 06:16:55 crc kubenswrapper[4818]: E1122 06:16:55.352496 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc0ca77c-a97b-467f-9c4d-a79612defabb" containerName="extract-content" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.352507 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc0ca77c-a97b-467f-9c4d-a79612defabb" containerName="extract-content" Nov 22 06:16:55 crc kubenswrapper[4818]: E1122 06:16:55.352556 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77eeb152-32d3-48f7-ba18-5b510b0bdd5c" containerName="collect-profiles" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.352566 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="77eeb152-32d3-48f7-ba18-5b510b0bdd5c" containerName="collect-profiles" Nov 22 06:16:55 crc kubenswrapper[4818]: E1122 06:16:55.352588 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cde37c03-b657-46b8-9a24-6ed81eac8c50" containerName="copy" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.352597 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="cde37c03-b657-46b8-9a24-6ed81eac8c50" containerName="copy" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.352885 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="cde37c03-b657-46b8-9a24-6ed81eac8c50" containerName="copy" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.352915 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc0ca77c-a97b-467f-9c4d-a79612defabb" containerName="registry-server" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.352940 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="cde37c03-b657-46b8-9a24-6ed81eac8c50" containerName="gather" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.352958 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="77eeb152-32d3-48f7-ba18-5b510b0bdd5c" containerName="collect-profiles" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.355588 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-drjjm" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.362994 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-drjjm"] Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.462537 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b567616e-5cd7-43f3-b51f-0c1d30b36be5-utilities\") pod \"redhat-marketplace-drjjm\" (UID: \"b567616e-5cd7-43f3-b51f-0c1d30b36be5\") " pod="openshift-marketplace/redhat-marketplace-drjjm" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.462658 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b567616e-5cd7-43f3-b51f-0c1d30b36be5-catalog-content\") pod \"redhat-marketplace-drjjm\" (UID: \"b567616e-5cd7-43f3-b51f-0c1d30b36be5\") " pod="openshift-marketplace/redhat-marketplace-drjjm" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.462693 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrctn\" (UniqueName: \"kubernetes.io/projected/b567616e-5cd7-43f3-b51f-0c1d30b36be5-kube-api-access-xrctn\") pod \"redhat-marketplace-drjjm\" (UID: \"b567616e-5cd7-43f3-b51f-0c1d30b36be5\") " pod="openshift-marketplace/redhat-marketplace-drjjm" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.564481 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b567616e-5cd7-43f3-b51f-0c1d30b36be5-utilities\") pod \"redhat-marketplace-drjjm\" (UID: \"b567616e-5cd7-43f3-b51f-0c1d30b36be5\") " pod="openshift-marketplace/redhat-marketplace-drjjm" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.564592 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b567616e-5cd7-43f3-b51f-0c1d30b36be5-catalog-content\") pod \"redhat-marketplace-drjjm\" (UID: \"b567616e-5cd7-43f3-b51f-0c1d30b36be5\") " pod="openshift-marketplace/redhat-marketplace-drjjm" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.564625 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrctn\" (UniqueName: \"kubernetes.io/projected/b567616e-5cd7-43f3-b51f-0c1d30b36be5-kube-api-access-xrctn\") pod \"redhat-marketplace-drjjm\" (UID: \"b567616e-5cd7-43f3-b51f-0c1d30b36be5\") " pod="openshift-marketplace/redhat-marketplace-drjjm" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.564992 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b567616e-5cd7-43f3-b51f-0c1d30b36be5-utilities\") pod \"redhat-marketplace-drjjm\" (UID: \"b567616e-5cd7-43f3-b51f-0c1d30b36be5\") " pod="openshift-marketplace/redhat-marketplace-drjjm" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.565238 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b567616e-5cd7-43f3-b51f-0c1d30b36be5-catalog-content\") pod \"redhat-marketplace-drjjm\" (UID: \"b567616e-5cd7-43f3-b51f-0c1d30b36be5\") " pod="openshift-marketplace/redhat-marketplace-drjjm" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.586361 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrctn\" (UniqueName: \"kubernetes.io/projected/b567616e-5cd7-43f3-b51f-0c1d30b36be5-kube-api-access-xrctn\") pod \"redhat-marketplace-drjjm\" (UID: \"b567616e-5cd7-43f3-b51f-0c1d30b36be5\") " pod="openshift-marketplace/redhat-marketplace-drjjm" Nov 22 06:16:55 crc kubenswrapper[4818]: I1122 06:16:55.709865 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-drjjm" Nov 22 06:16:56 crc kubenswrapper[4818]: I1122 06:16:56.203531 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-drjjm"] Nov 22 06:16:56 crc kubenswrapper[4818]: I1122 06:16:56.298526 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:16:56 crc kubenswrapper[4818]: E1122 06:16:56.299048 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:16:56 crc kubenswrapper[4818]: I1122 06:16:56.554600 4818 generic.go:334] "Generic (PLEG): container finished" podID="b567616e-5cd7-43f3-b51f-0c1d30b36be5" containerID="cc24a97381a10120c27c501c2909348c63e8e7a8472841b23161d54fb7a0859a" exitCode=0 Nov 22 06:16:56 crc kubenswrapper[4818]: I1122 06:16:56.554662 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-drjjm" event={"ID":"b567616e-5cd7-43f3-b51f-0c1d30b36be5","Type":"ContainerDied","Data":"cc24a97381a10120c27c501c2909348c63e8e7a8472841b23161d54fb7a0859a"} Nov 22 06:16:56 crc kubenswrapper[4818]: I1122 06:16:56.554699 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-drjjm" event={"ID":"b567616e-5cd7-43f3-b51f-0c1d30b36be5","Type":"ContainerStarted","Data":"0eeb5ad863480d03b1634dfd8a16a11bb11a8fb902737c142ef0fe2965217213"} Nov 22 06:16:58 crc kubenswrapper[4818]: I1122 06:16:58.579351 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-drjjm" event={"ID":"b567616e-5cd7-43f3-b51f-0c1d30b36be5","Type":"ContainerStarted","Data":"fd19d1aa188f583ef809279fc2afc9fb77899d3ff6b2befa8141d0cc0454c2e7"} Nov 22 06:16:59 crc kubenswrapper[4818]: I1122 06:16:59.594918 4818 generic.go:334] "Generic (PLEG): container finished" podID="b567616e-5cd7-43f3-b51f-0c1d30b36be5" containerID="fd19d1aa188f583ef809279fc2afc9fb77899d3ff6b2befa8141d0cc0454c2e7" exitCode=0 Nov 22 06:16:59 crc kubenswrapper[4818]: I1122 06:16:59.595142 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-drjjm" event={"ID":"b567616e-5cd7-43f3-b51f-0c1d30b36be5","Type":"ContainerDied","Data":"fd19d1aa188f583ef809279fc2afc9fb77899d3ff6b2befa8141d0cc0454c2e7"} Nov 22 06:17:00 crc kubenswrapper[4818]: I1122 06:17:00.608077 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-drjjm" event={"ID":"b567616e-5cd7-43f3-b51f-0c1d30b36be5","Type":"ContainerStarted","Data":"dbee5fddcbe488f87f7c7836d94fdd8bae2e9d7fb8d45325088895c42e18aaa4"} Nov 22 06:17:00 crc kubenswrapper[4818]: I1122 06:17:00.635758 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-drjjm" podStartSLOduration=2.138468664 podStartE2EDuration="5.635732796s" podCreationTimestamp="2025-11-22 06:16:55 +0000 UTC" firstStartedPulling="2025-11-22 06:16:56.557715028 +0000 UTC m=+5369.132131555" lastFinishedPulling="2025-11-22 06:17:00.05497916 +0000 UTC m=+5372.629395687" observedRunningTime="2025-11-22 06:17:00.629373246 +0000 UTC m=+5373.203789773" watchObservedRunningTime="2025-11-22 06:17:00.635732796 +0000 UTC m=+5373.210149323" Nov 22 06:17:05 crc kubenswrapper[4818]: I1122 06:17:05.710502 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-drjjm" Nov 22 06:17:05 crc kubenswrapper[4818]: I1122 06:17:05.711359 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-drjjm" Nov 22 06:17:05 crc kubenswrapper[4818]: I1122 06:17:05.991244 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-drjjm" Nov 22 06:17:06 crc kubenswrapper[4818]: I1122 06:17:06.739165 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-drjjm" Nov 22 06:17:06 crc kubenswrapper[4818]: I1122 06:17:06.792057 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-drjjm"] Nov 22 06:17:08 crc kubenswrapper[4818]: I1122 06:17:08.303383 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:17:08 crc kubenswrapper[4818]: E1122 06:17:08.304114 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:17:08 crc kubenswrapper[4818]: I1122 06:17:08.699799 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-drjjm" podUID="b567616e-5cd7-43f3-b51f-0c1d30b36be5" containerName="registry-server" containerID="cri-o://dbee5fddcbe488f87f7c7836d94fdd8bae2e9d7fb8d45325088895c42e18aaa4" gracePeriod=2 Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.162948 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-drjjm" Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.204416 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b567616e-5cd7-43f3-b51f-0c1d30b36be5-catalog-content\") pod \"b567616e-5cd7-43f3-b51f-0c1d30b36be5\" (UID: \"b567616e-5cd7-43f3-b51f-0c1d30b36be5\") " Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.204471 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrctn\" (UniqueName: \"kubernetes.io/projected/b567616e-5cd7-43f3-b51f-0c1d30b36be5-kube-api-access-xrctn\") pod \"b567616e-5cd7-43f3-b51f-0c1d30b36be5\" (UID: \"b567616e-5cd7-43f3-b51f-0c1d30b36be5\") " Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.204511 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b567616e-5cd7-43f3-b51f-0c1d30b36be5-utilities\") pod \"b567616e-5cd7-43f3-b51f-0c1d30b36be5\" (UID: \"b567616e-5cd7-43f3-b51f-0c1d30b36be5\") " Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.205984 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b567616e-5cd7-43f3-b51f-0c1d30b36be5-utilities" (OuterVolumeSpecName: "utilities") pod "b567616e-5cd7-43f3-b51f-0c1d30b36be5" (UID: "b567616e-5cd7-43f3-b51f-0c1d30b36be5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.210528 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b567616e-5cd7-43f3-b51f-0c1d30b36be5-kube-api-access-xrctn" (OuterVolumeSpecName: "kube-api-access-xrctn") pod "b567616e-5cd7-43f3-b51f-0c1d30b36be5" (UID: "b567616e-5cd7-43f3-b51f-0c1d30b36be5"). InnerVolumeSpecName "kube-api-access-xrctn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.241162 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b567616e-5cd7-43f3-b51f-0c1d30b36be5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b567616e-5cd7-43f3-b51f-0c1d30b36be5" (UID: "b567616e-5cd7-43f3-b51f-0c1d30b36be5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.306588 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrctn\" (UniqueName: \"kubernetes.io/projected/b567616e-5cd7-43f3-b51f-0c1d30b36be5-kube-api-access-xrctn\") on node \"crc\" DevicePath \"\"" Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.306634 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b567616e-5cd7-43f3-b51f-0c1d30b36be5-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.306646 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b567616e-5cd7-43f3-b51f-0c1d30b36be5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.713621 4818 generic.go:334] "Generic (PLEG): container finished" podID="b567616e-5cd7-43f3-b51f-0c1d30b36be5" containerID="dbee5fddcbe488f87f7c7836d94fdd8bae2e9d7fb8d45325088895c42e18aaa4" exitCode=0 Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.714031 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-drjjm" event={"ID":"b567616e-5cd7-43f3-b51f-0c1d30b36be5","Type":"ContainerDied","Data":"dbee5fddcbe488f87f7c7836d94fdd8bae2e9d7fb8d45325088895c42e18aaa4"} Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.714078 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-drjjm" event={"ID":"b567616e-5cd7-43f3-b51f-0c1d30b36be5","Type":"ContainerDied","Data":"0eeb5ad863480d03b1634dfd8a16a11bb11a8fb902737c142ef0fe2965217213"} Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.714109 4818 scope.go:117] "RemoveContainer" containerID="dbee5fddcbe488f87f7c7836d94fdd8bae2e9d7fb8d45325088895c42e18aaa4" Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.714366 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-drjjm" Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.777059 4818 scope.go:117] "RemoveContainer" containerID="fd19d1aa188f583ef809279fc2afc9fb77899d3ff6b2befa8141d0cc0454c2e7" Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.783808 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-drjjm"] Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.798629 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-drjjm"] Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.816040 4818 scope.go:117] "RemoveContainer" containerID="cc24a97381a10120c27c501c2909348c63e8e7a8472841b23161d54fb7a0859a" Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.875936 4818 scope.go:117] "RemoveContainer" containerID="dbee5fddcbe488f87f7c7836d94fdd8bae2e9d7fb8d45325088895c42e18aaa4" Nov 22 06:17:09 crc kubenswrapper[4818]: E1122 06:17:09.876638 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbee5fddcbe488f87f7c7836d94fdd8bae2e9d7fb8d45325088895c42e18aaa4\": container with ID starting with dbee5fddcbe488f87f7c7836d94fdd8bae2e9d7fb8d45325088895c42e18aaa4 not found: ID does not exist" containerID="dbee5fddcbe488f87f7c7836d94fdd8bae2e9d7fb8d45325088895c42e18aaa4" Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.876840 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbee5fddcbe488f87f7c7836d94fdd8bae2e9d7fb8d45325088895c42e18aaa4"} err="failed to get container status \"dbee5fddcbe488f87f7c7836d94fdd8bae2e9d7fb8d45325088895c42e18aaa4\": rpc error: code = NotFound desc = could not find container \"dbee5fddcbe488f87f7c7836d94fdd8bae2e9d7fb8d45325088895c42e18aaa4\": container with ID starting with dbee5fddcbe488f87f7c7836d94fdd8bae2e9d7fb8d45325088895c42e18aaa4 not found: ID does not exist" Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.876984 4818 scope.go:117] "RemoveContainer" containerID="fd19d1aa188f583ef809279fc2afc9fb77899d3ff6b2befa8141d0cc0454c2e7" Nov 22 06:17:09 crc kubenswrapper[4818]: E1122 06:17:09.877687 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd19d1aa188f583ef809279fc2afc9fb77899d3ff6b2befa8141d0cc0454c2e7\": container with ID starting with fd19d1aa188f583ef809279fc2afc9fb77899d3ff6b2befa8141d0cc0454c2e7 not found: ID does not exist" containerID="fd19d1aa188f583ef809279fc2afc9fb77899d3ff6b2befa8141d0cc0454c2e7" Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.877741 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd19d1aa188f583ef809279fc2afc9fb77899d3ff6b2befa8141d0cc0454c2e7"} err="failed to get container status \"fd19d1aa188f583ef809279fc2afc9fb77899d3ff6b2befa8141d0cc0454c2e7\": rpc error: code = NotFound desc = could not find container \"fd19d1aa188f583ef809279fc2afc9fb77899d3ff6b2befa8141d0cc0454c2e7\": container with ID starting with fd19d1aa188f583ef809279fc2afc9fb77899d3ff6b2befa8141d0cc0454c2e7 not found: ID does not exist" Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.877807 4818 scope.go:117] "RemoveContainer" containerID="cc24a97381a10120c27c501c2909348c63e8e7a8472841b23161d54fb7a0859a" Nov 22 06:17:09 crc kubenswrapper[4818]: E1122 06:17:09.878187 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc24a97381a10120c27c501c2909348c63e8e7a8472841b23161d54fb7a0859a\": container with ID starting with cc24a97381a10120c27c501c2909348c63e8e7a8472841b23161d54fb7a0859a not found: ID does not exist" containerID="cc24a97381a10120c27c501c2909348c63e8e7a8472841b23161d54fb7a0859a" Nov 22 06:17:09 crc kubenswrapper[4818]: I1122 06:17:09.878348 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc24a97381a10120c27c501c2909348c63e8e7a8472841b23161d54fb7a0859a"} err="failed to get container status \"cc24a97381a10120c27c501c2909348c63e8e7a8472841b23161d54fb7a0859a\": rpc error: code = NotFound desc = could not find container \"cc24a97381a10120c27c501c2909348c63e8e7a8472841b23161d54fb7a0859a\": container with ID starting with cc24a97381a10120c27c501c2909348c63e8e7a8472841b23161d54fb7a0859a not found: ID does not exist" Nov 22 06:17:10 crc kubenswrapper[4818]: I1122 06:17:10.303555 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b567616e-5cd7-43f3-b51f-0c1d30b36be5" path="/var/lib/kubelet/pods/b567616e-5cd7-43f3-b51f-0c1d30b36be5/volumes" Nov 22 06:17:22 crc kubenswrapper[4818]: I1122 06:17:22.290906 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:17:22 crc kubenswrapper[4818]: E1122 06:17:22.291868 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:17:34 crc kubenswrapper[4818]: I1122 06:17:34.292102 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:17:34 crc kubenswrapper[4818]: E1122 06:17:34.293564 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:17:38 crc kubenswrapper[4818]: I1122 06:17:38.840692 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2rzhz"] Nov 22 06:17:38 crc kubenswrapper[4818]: E1122 06:17:38.845323 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b567616e-5cd7-43f3-b51f-0c1d30b36be5" containerName="extract-content" Nov 22 06:17:38 crc kubenswrapper[4818]: I1122 06:17:38.845365 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b567616e-5cd7-43f3-b51f-0c1d30b36be5" containerName="extract-content" Nov 22 06:17:38 crc kubenswrapper[4818]: E1122 06:17:38.851189 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b567616e-5cd7-43f3-b51f-0c1d30b36be5" containerName="registry-server" Nov 22 06:17:38 crc kubenswrapper[4818]: I1122 06:17:38.851239 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b567616e-5cd7-43f3-b51f-0c1d30b36be5" containerName="registry-server" Nov 22 06:17:38 crc kubenswrapper[4818]: E1122 06:17:38.851404 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b567616e-5cd7-43f3-b51f-0c1d30b36be5" containerName="extract-utilities" Nov 22 06:17:38 crc kubenswrapper[4818]: I1122 06:17:38.851418 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b567616e-5cd7-43f3-b51f-0c1d30b36be5" containerName="extract-utilities" Nov 22 06:17:38 crc kubenswrapper[4818]: I1122 06:17:38.852226 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="b567616e-5cd7-43f3-b51f-0c1d30b36be5" containerName="registry-server" Nov 22 06:17:38 crc kubenswrapper[4818]: I1122 06:17:38.856346 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2rzhz" Nov 22 06:17:38 crc kubenswrapper[4818]: I1122 06:17:38.887400 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2rzhz"] Nov 22 06:17:38 crc kubenswrapper[4818]: I1122 06:17:38.904317 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ec5ccb2-0e1d-459d-98a8-dceceb965ee7-catalog-content\") pod \"certified-operators-2rzhz\" (UID: \"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7\") " pod="openshift-marketplace/certified-operators-2rzhz" Nov 22 06:17:38 crc kubenswrapper[4818]: I1122 06:17:38.904427 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5qqv\" (UniqueName: \"kubernetes.io/projected/1ec5ccb2-0e1d-459d-98a8-dceceb965ee7-kube-api-access-k5qqv\") pod \"certified-operators-2rzhz\" (UID: \"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7\") " pod="openshift-marketplace/certified-operators-2rzhz" Nov 22 06:17:38 crc kubenswrapper[4818]: I1122 06:17:38.904488 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ec5ccb2-0e1d-459d-98a8-dceceb965ee7-utilities\") pod \"certified-operators-2rzhz\" (UID: \"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7\") " pod="openshift-marketplace/certified-operators-2rzhz" Nov 22 06:17:39 crc kubenswrapper[4818]: I1122 06:17:39.006406 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ec5ccb2-0e1d-459d-98a8-dceceb965ee7-catalog-content\") pod \"certified-operators-2rzhz\" (UID: \"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7\") " pod="openshift-marketplace/certified-operators-2rzhz" Nov 22 06:17:39 crc kubenswrapper[4818]: I1122 06:17:39.006691 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5qqv\" (UniqueName: \"kubernetes.io/projected/1ec5ccb2-0e1d-459d-98a8-dceceb965ee7-kube-api-access-k5qqv\") pod \"certified-operators-2rzhz\" (UID: \"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7\") " pod="openshift-marketplace/certified-operators-2rzhz" Nov 22 06:17:39 crc kubenswrapper[4818]: I1122 06:17:39.006745 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ec5ccb2-0e1d-459d-98a8-dceceb965ee7-utilities\") pod \"certified-operators-2rzhz\" (UID: \"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7\") " pod="openshift-marketplace/certified-operators-2rzhz" Nov 22 06:17:39 crc kubenswrapper[4818]: I1122 06:17:39.007206 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ec5ccb2-0e1d-459d-98a8-dceceb965ee7-utilities\") pod \"certified-operators-2rzhz\" (UID: \"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7\") " pod="openshift-marketplace/certified-operators-2rzhz" Nov 22 06:17:39 crc kubenswrapper[4818]: I1122 06:17:39.007208 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ec5ccb2-0e1d-459d-98a8-dceceb965ee7-catalog-content\") pod \"certified-operators-2rzhz\" (UID: \"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7\") " pod="openshift-marketplace/certified-operators-2rzhz" Nov 22 06:17:39 crc kubenswrapper[4818]: I1122 06:17:39.027994 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5qqv\" (UniqueName: \"kubernetes.io/projected/1ec5ccb2-0e1d-459d-98a8-dceceb965ee7-kube-api-access-k5qqv\") pod \"certified-operators-2rzhz\" (UID: \"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7\") " pod="openshift-marketplace/certified-operators-2rzhz" Nov 22 06:17:39 crc kubenswrapper[4818]: I1122 06:17:39.198605 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2rzhz" Nov 22 06:17:39 crc kubenswrapper[4818]: I1122 06:17:39.743286 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2rzhz"] Nov 22 06:17:40 crc kubenswrapper[4818]: I1122 06:17:40.042378 4818 generic.go:334] "Generic (PLEG): container finished" podID="1ec5ccb2-0e1d-459d-98a8-dceceb965ee7" containerID="1b84055d84b6022de7c6b190143cd185a7203bc244c51be83d34501d0c2a56e4" exitCode=0 Nov 22 06:17:40 crc kubenswrapper[4818]: I1122 06:17:40.042484 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2rzhz" event={"ID":"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7","Type":"ContainerDied","Data":"1b84055d84b6022de7c6b190143cd185a7203bc244c51be83d34501d0c2a56e4"} Nov 22 06:17:40 crc kubenswrapper[4818]: I1122 06:17:40.042679 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2rzhz" event={"ID":"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7","Type":"ContainerStarted","Data":"664d60ac336f4a873c9645407234004e2b943176af7c7aabf40d52b3a3cb5e03"} Nov 22 06:17:41 crc kubenswrapper[4818]: I1122 06:17:41.050687 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2rzhz" event={"ID":"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7","Type":"ContainerStarted","Data":"1738c2eafab21755ea4e952d8f3eb47002e7263a933d92321cebf15c60fa1155"} Nov 22 06:17:44 crc kubenswrapper[4818]: I1122 06:17:44.081356 4818 generic.go:334] "Generic (PLEG): container finished" podID="1ec5ccb2-0e1d-459d-98a8-dceceb965ee7" containerID="1738c2eafab21755ea4e952d8f3eb47002e7263a933d92321cebf15c60fa1155" exitCode=0 Nov 22 06:17:44 crc kubenswrapper[4818]: I1122 06:17:44.081555 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2rzhz" event={"ID":"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7","Type":"ContainerDied","Data":"1738c2eafab21755ea4e952d8f3eb47002e7263a933d92321cebf15c60fa1155"} Nov 22 06:17:45 crc kubenswrapper[4818]: I1122 06:17:45.094875 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2rzhz" event={"ID":"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7","Type":"ContainerStarted","Data":"de1f8238822a78ca1b4a319d401259343d6200739f40ad185a7d6aac3d9ca3a4"} Nov 22 06:17:45 crc kubenswrapper[4818]: I1122 06:17:45.154607 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2rzhz" podStartSLOduration=2.708691707 podStartE2EDuration="7.154586526s" podCreationTimestamp="2025-11-22 06:17:38 +0000 UTC" firstStartedPulling="2025-11-22 06:17:40.044320963 +0000 UTC m=+5412.618737490" lastFinishedPulling="2025-11-22 06:17:44.490215752 +0000 UTC m=+5417.064632309" observedRunningTime="2025-11-22 06:17:45.141471015 +0000 UTC m=+5417.715887552" watchObservedRunningTime="2025-11-22 06:17:45.154586526 +0000 UTC m=+5417.729003063" Nov 22 06:17:46 crc kubenswrapper[4818]: I1122 06:17:46.292021 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:17:46 crc kubenswrapper[4818]: E1122 06:17:46.292705 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:17:49 crc kubenswrapper[4818]: I1122 06:17:49.199659 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2rzhz" Nov 22 06:17:49 crc kubenswrapper[4818]: I1122 06:17:49.200022 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2rzhz" Nov 22 06:17:49 crc kubenswrapper[4818]: I1122 06:17:49.285283 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2rzhz" Nov 22 06:17:50 crc kubenswrapper[4818]: I1122 06:17:50.194045 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2rzhz" Nov 22 06:17:50 crc kubenswrapper[4818]: I1122 06:17:50.257044 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2rzhz"] Nov 22 06:17:52 crc kubenswrapper[4818]: I1122 06:17:52.165188 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2rzhz" podUID="1ec5ccb2-0e1d-459d-98a8-dceceb965ee7" containerName="registry-server" containerID="cri-o://de1f8238822a78ca1b4a319d401259343d6200739f40ad185a7d6aac3d9ca3a4" gracePeriod=2 Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.117371 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2rzhz" Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.176016 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ec5ccb2-0e1d-459d-98a8-dceceb965ee7-utilities\") pod \"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7\" (UID: \"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7\") " Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.176365 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5qqv\" (UniqueName: \"kubernetes.io/projected/1ec5ccb2-0e1d-459d-98a8-dceceb965ee7-kube-api-access-k5qqv\") pod \"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7\" (UID: \"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7\") " Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.176411 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ec5ccb2-0e1d-459d-98a8-dceceb965ee7-catalog-content\") pod \"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7\" (UID: \"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7\") " Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.176965 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ec5ccb2-0e1d-459d-98a8-dceceb965ee7-utilities" (OuterVolumeSpecName: "utilities") pod "1ec5ccb2-0e1d-459d-98a8-dceceb965ee7" (UID: "1ec5ccb2-0e1d-459d-98a8-dceceb965ee7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.181594 4818 generic.go:334] "Generic (PLEG): container finished" podID="1ec5ccb2-0e1d-459d-98a8-dceceb965ee7" containerID="de1f8238822a78ca1b4a319d401259343d6200739f40ad185a7d6aac3d9ca3a4" exitCode=0 Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.181649 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2rzhz" event={"ID":"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7","Type":"ContainerDied","Data":"de1f8238822a78ca1b4a319d401259343d6200739f40ad185a7d6aac3d9ca3a4"} Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.181663 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2rzhz" Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.181679 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2rzhz" event={"ID":"1ec5ccb2-0e1d-459d-98a8-dceceb965ee7","Type":"ContainerDied","Data":"664d60ac336f4a873c9645407234004e2b943176af7c7aabf40d52b3a3cb5e03"} Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.181698 4818 scope.go:117] "RemoveContainer" containerID="de1f8238822a78ca1b4a319d401259343d6200739f40ad185a7d6aac3d9ca3a4" Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.197232 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ec5ccb2-0e1d-459d-98a8-dceceb965ee7-kube-api-access-k5qqv" (OuterVolumeSpecName: "kube-api-access-k5qqv") pod "1ec5ccb2-0e1d-459d-98a8-dceceb965ee7" (UID: "1ec5ccb2-0e1d-459d-98a8-dceceb965ee7"). InnerVolumeSpecName "kube-api-access-k5qqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.238736 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ec5ccb2-0e1d-459d-98a8-dceceb965ee7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ec5ccb2-0e1d-459d-98a8-dceceb965ee7" (UID: "1ec5ccb2-0e1d-459d-98a8-dceceb965ee7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.240239 4818 scope.go:117] "RemoveContainer" containerID="1738c2eafab21755ea4e952d8f3eb47002e7263a933d92321cebf15c60fa1155" Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.258805 4818 scope.go:117] "RemoveContainer" containerID="1b84055d84b6022de7c6b190143cd185a7203bc244c51be83d34501d0c2a56e4" Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.280605 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ec5ccb2-0e1d-459d-98a8-dceceb965ee7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.280990 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ec5ccb2-0e1d-459d-98a8-dceceb965ee7-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.281210 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5qqv\" (UniqueName: \"kubernetes.io/projected/1ec5ccb2-0e1d-459d-98a8-dceceb965ee7-kube-api-access-k5qqv\") on node \"crc\" DevicePath \"\"" Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.304319 4818 scope.go:117] "RemoveContainer" containerID="de1f8238822a78ca1b4a319d401259343d6200739f40ad185a7d6aac3d9ca3a4" Nov 22 06:17:53 crc kubenswrapper[4818]: E1122 06:17:53.305014 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de1f8238822a78ca1b4a319d401259343d6200739f40ad185a7d6aac3d9ca3a4\": container with ID starting with de1f8238822a78ca1b4a319d401259343d6200739f40ad185a7d6aac3d9ca3a4 not found: ID does not exist" containerID="de1f8238822a78ca1b4a319d401259343d6200739f40ad185a7d6aac3d9ca3a4" Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.305062 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de1f8238822a78ca1b4a319d401259343d6200739f40ad185a7d6aac3d9ca3a4"} err="failed to get container status \"de1f8238822a78ca1b4a319d401259343d6200739f40ad185a7d6aac3d9ca3a4\": rpc error: code = NotFound desc = could not find container \"de1f8238822a78ca1b4a319d401259343d6200739f40ad185a7d6aac3d9ca3a4\": container with ID starting with de1f8238822a78ca1b4a319d401259343d6200739f40ad185a7d6aac3d9ca3a4 not found: ID does not exist" Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.305087 4818 scope.go:117] "RemoveContainer" containerID="1738c2eafab21755ea4e952d8f3eb47002e7263a933d92321cebf15c60fa1155" Nov 22 06:17:53 crc kubenswrapper[4818]: E1122 06:17:53.305478 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1738c2eafab21755ea4e952d8f3eb47002e7263a933d92321cebf15c60fa1155\": container with ID starting with 1738c2eafab21755ea4e952d8f3eb47002e7263a933d92321cebf15c60fa1155 not found: ID does not exist" containerID="1738c2eafab21755ea4e952d8f3eb47002e7263a933d92321cebf15c60fa1155" Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.305749 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1738c2eafab21755ea4e952d8f3eb47002e7263a933d92321cebf15c60fa1155"} err="failed to get container status \"1738c2eafab21755ea4e952d8f3eb47002e7263a933d92321cebf15c60fa1155\": rpc error: code = NotFound desc = could not find container \"1738c2eafab21755ea4e952d8f3eb47002e7263a933d92321cebf15c60fa1155\": container with ID starting with 1738c2eafab21755ea4e952d8f3eb47002e7263a933d92321cebf15c60fa1155 not found: ID does not exist" Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.305790 4818 scope.go:117] "RemoveContainer" containerID="1b84055d84b6022de7c6b190143cd185a7203bc244c51be83d34501d0c2a56e4" Nov 22 06:17:53 crc kubenswrapper[4818]: E1122 06:17:53.306090 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b84055d84b6022de7c6b190143cd185a7203bc244c51be83d34501d0c2a56e4\": container with ID starting with 1b84055d84b6022de7c6b190143cd185a7203bc244c51be83d34501d0c2a56e4 not found: ID does not exist" containerID="1b84055d84b6022de7c6b190143cd185a7203bc244c51be83d34501d0c2a56e4" Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.306127 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b84055d84b6022de7c6b190143cd185a7203bc244c51be83d34501d0c2a56e4"} err="failed to get container status \"1b84055d84b6022de7c6b190143cd185a7203bc244c51be83d34501d0c2a56e4\": rpc error: code = NotFound desc = could not find container \"1b84055d84b6022de7c6b190143cd185a7203bc244c51be83d34501d0c2a56e4\": container with ID starting with 1b84055d84b6022de7c6b190143cd185a7203bc244c51be83d34501d0c2a56e4 not found: ID does not exist" Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.532351 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2rzhz"] Nov 22 06:17:53 crc kubenswrapper[4818]: I1122 06:17:53.542199 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2rzhz"] Nov 22 06:17:54 crc kubenswrapper[4818]: I1122 06:17:54.302225 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ec5ccb2-0e1d-459d-98a8-dceceb965ee7" path="/var/lib/kubelet/pods/1ec5ccb2-0e1d-459d-98a8-dceceb965ee7/volumes" Nov 22 06:17:59 crc kubenswrapper[4818]: I1122 06:17:59.292210 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:17:59 crc kubenswrapper[4818]: E1122 06:17:59.293649 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:18:13 crc kubenswrapper[4818]: I1122 06:18:13.292335 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:18:13 crc kubenswrapper[4818]: E1122 06:18:13.293607 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rjvw_openshift-machine-config-operator(0971c9ec-29ce-44eb-9875-ad4664143305)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" podUID="0971c9ec-29ce-44eb-9875-ad4664143305" Nov 22 06:18:24 crc kubenswrapper[4818]: I1122 06:18:24.292558 4818 scope.go:117] "RemoveContainer" containerID="07359eab392ca16828241b4025b1b8916161d0e37089f0ddb34843140ac544fa" Nov 22 06:18:25 crc kubenswrapper[4818]: I1122 06:18:25.609008 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rjvw" event={"ID":"0971c9ec-29ce-44eb-9875-ad4664143305","Type":"ContainerStarted","Data":"2fe21199584fc09bc0978bce6f6c96dd6108b964f8cb05a8d39cd2a0452ab1ea"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515110252761024445 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015110252761017362 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015110237560016505 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015110237560015455 5ustar corecore